var/home/core/zuul-output/0000755000175000017500000000000015111401654014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111406621015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003607205315111406612017700 0ustar rootrootNov 25 19:33:28 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 19:33:28 crc restorecon[4710]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:28 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 19:33:29 crc restorecon[4710]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 19:33:29 crc kubenswrapper[4759]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.838610 4759 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.849516 4759 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.849829 4759 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.849948 4759 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850530 4759 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850590 4759 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850605 4759 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850616 4759 feature_gate.go:330] unrecognized feature gate: Example Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850627 4759 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850638 4759 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850650 4759 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850660 4759 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850677 4759 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850698 4759 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850710 4759 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850723 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850734 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850745 4759 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850756 4759 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850766 4759 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850776 4759 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850787 4759 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850796 4759 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850804 4759 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850812 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850820 4759 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850845 4759 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850854 4759 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850862 4759 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850871 4759 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850880 4759 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850888 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850896 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850904 4759 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850918 4759 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850926 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850935 4759 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850945 4759 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850952 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850960 4759 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850969 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850980 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850990 4759 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.850998 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851006 4759 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851014 4759 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851021 4759 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851029 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851041 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851048 4759 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851056 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851064 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851073 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851080 4759 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851088 4759 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851095 4759 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851104 4759 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851112 4759 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851120 4759 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851127 4759 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851135 4759 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851146 4759 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851158 4759 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851168 4759 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851178 4759 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851187 4759 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851199 4759 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851209 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851220 4759 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851228 4759 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851246 4759 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.851257 4759 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851510 4759 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851532 4759 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851549 4759 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851561 4759 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851573 4759 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851583 4759 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851595 4759 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851607 4759 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851616 4759 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851625 4759 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851635 4759 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851649 4759 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851658 4759 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851667 4759 flags.go:64] FLAG: --cgroup-root="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851676 4759 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851685 4759 flags.go:64] FLAG: --client-ca-file="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851694 4759 flags.go:64] FLAG: --cloud-config="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851703 4759 flags.go:64] FLAG: --cloud-provider="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851712 4759 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851724 4759 flags.go:64] FLAG: --cluster-domain="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851733 4759 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851743 4759 flags.go:64] FLAG: --config-dir="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851752 4759 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851762 4759 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851773 4759 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851782 4759 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851791 4759 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851803 4759 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851816 4759 flags.go:64] FLAG: --contention-profiling="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851827 4759 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851838 4759 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851850 4759 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851862 4759 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851878 4759 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851890 4759 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851902 4759 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851914 4759 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851928 4759 flags.go:64] FLAG: --enable-server="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851940 4759 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851956 4759 flags.go:64] FLAG: --event-burst="100" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851970 4759 flags.go:64] FLAG: --event-qps="50" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851982 4759 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.851995 4759 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852006 4759 flags.go:64] FLAG: --eviction-hard="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852017 4759 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852028 4759 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852037 4759 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852048 4759 flags.go:64] FLAG: --eviction-soft="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852058 4759 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852066 4759 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852076 4759 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852084 4759 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852177 4759 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852188 4759 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852196 4759 flags.go:64] FLAG: --feature-gates="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852208 4759 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852217 4759 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852227 4759 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852236 4759 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852246 4759 flags.go:64] FLAG: --healthz-port="10248" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852256 4759 flags.go:64] FLAG: --help="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852265 4759 flags.go:64] FLAG: --hostname-override="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852273 4759 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852283 4759 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852338 4759 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852348 4759 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852357 4759 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852366 4759 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852375 4759 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852384 4759 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852393 4759 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852402 4759 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852412 4759 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852421 4759 flags.go:64] FLAG: --kube-reserved="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852431 4759 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852440 4759 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852477 4759 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852486 4759 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852495 4759 flags.go:64] FLAG: --lock-file="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852506 4759 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852515 4759 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852525 4759 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852540 4759 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852551 4759 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852561 4759 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852570 4759 flags.go:64] FLAG: --logging-format="text" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852579 4759 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852588 4759 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852597 4759 flags.go:64] FLAG: --manifest-url="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852606 4759 flags.go:64] FLAG: --manifest-url-header="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852619 4759 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852628 4759 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852639 4759 flags.go:64] FLAG: --max-pods="110" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852648 4759 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852657 4759 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852667 4759 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852675 4759 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852685 4759 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852694 4759 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852703 4759 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852725 4759 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852734 4759 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852745 4759 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852754 4759 flags.go:64] FLAG: --pod-cidr="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852763 4759 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852778 4759 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852787 4759 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852796 4759 flags.go:64] FLAG: --pods-per-core="0" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852805 4759 flags.go:64] FLAG: --port="10250" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852814 4759 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852824 4759 flags.go:64] FLAG: --provider-id="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852833 4759 flags.go:64] FLAG: --qos-reserved="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852842 4759 flags.go:64] FLAG: --read-only-port="10255" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852851 4759 flags.go:64] FLAG: --register-node="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852860 4759 flags.go:64] FLAG: --register-schedulable="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852869 4759 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852935 4759 flags.go:64] FLAG: --registry-burst="10" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852944 4759 flags.go:64] FLAG: --registry-qps="5" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852953 4759 flags.go:64] FLAG: --reserved-cpus="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852963 4759 flags.go:64] FLAG: --reserved-memory="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852975 4759 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852984 4759 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.852996 4759 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853007 4759 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853026 4759 flags.go:64] FLAG: --runonce="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853038 4759 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853050 4759 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853062 4759 flags.go:64] FLAG: --seccomp-default="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853074 4759 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853087 4759 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853100 4759 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853114 4759 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853126 4759 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853137 4759 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853149 4759 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853161 4759 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853173 4759 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853186 4759 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853198 4759 flags.go:64] FLAG: --system-cgroups="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853210 4759 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853230 4759 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853239 4759 flags.go:64] FLAG: --tls-cert-file="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853249 4759 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853262 4759 flags.go:64] FLAG: --tls-min-version="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853271 4759 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853280 4759 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853289 4759 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853298 4759 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853307 4759 flags.go:64] FLAG: --v="2" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853320 4759 flags.go:64] FLAG: --version="false" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853332 4759 flags.go:64] FLAG: --vmodule="" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853343 4759 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.853353 4759 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853650 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853662 4759 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853673 4759 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853686 4759 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853695 4759 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853702 4759 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853710 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853718 4759 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853726 4759 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853734 4759 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853742 4759 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853750 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853758 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853765 4759 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853777 4759 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853786 4759 feature_gate.go:330] unrecognized feature gate: Example Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853796 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853805 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853813 4759 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853822 4759 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853831 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853840 4759 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853847 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853855 4759 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853863 4759 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853871 4759 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853879 4759 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853886 4759 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853894 4759 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853902 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853910 4759 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853917 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853925 4759 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853933 4759 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853940 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853951 4759 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853961 4759 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853970 4759 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853979 4759 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853987 4759 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.853995 4759 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854003 4759 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854013 4759 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854022 4759 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854030 4759 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854038 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854047 4759 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854055 4759 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854064 4759 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854073 4759 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854081 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854090 4759 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854099 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854110 4759 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854118 4759 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854126 4759 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854134 4759 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854142 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854150 4759 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854159 4759 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854167 4759 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854176 4759 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854186 4759 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854196 4759 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854204 4759 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854213 4759 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854221 4759 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854232 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854241 4759 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854252 4759 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.854263 4759 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.854279 4759 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.866148 4759 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.866191 4759 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866372 4759 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866392 4759 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866405 4759 feature_gate.go:330] unrecognized feature gate: Example Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866417 4759 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866427 4759 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866437 4759 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866483 4759 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866494 4759 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866504 4759 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866514 4759 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866528 4759 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866544 4759 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866555 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866566 4759 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866579 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866589 4759 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866602 4759 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866615 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866626 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866637 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866648 4759 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866659 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866668 4759 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866678 4759 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866689 4759 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866699 4759 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866709 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866720 4759 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866731 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866742 4759 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866752 4759 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866762 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866772 4759 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866782 4759 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866796 4759 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866806 4759 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866817 4759 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866827 4759 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866836 4759 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866847 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866856 4759 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866866 4759 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866877 4759 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866886 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866897 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866907 4759 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866917 4759 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866926 4759 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866937 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866948 4759 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866959 4759 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866969 4759 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866980 4759 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.866990 4759 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867002 4759 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867016 4759 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867030 4759 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867041 4759 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867053 4759 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867065 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867077 4759 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867088 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867101 4759 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867113 4759 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867123 4759 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867137 4759 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867149 4759 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867159 4759 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867169 4759 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867180 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867193 4759 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.867211 4759 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867545 4759 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867565 4759 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867578 4759 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867591 4759 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867601 4759 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867611 4759 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867621 4759 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867633 4759 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867643 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867653 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867662 4759 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867672 4759 feature_gate.go:330] unrecognized feature gate: Example Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867682 4759 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867692 4759 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867707 4759 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867722 4759 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867734 4759 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867744 4759 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867754 4759 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867765 4759 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867775 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867786 4759 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867796 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867806 4759 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867817 4759 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867827 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867837 4759 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867847 4759 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867857 4759 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867868 4759 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867878 4759 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867889 4759 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867898 4759 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867908 4759 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867921 4759 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867932 4759 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867942 4759 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867952 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867966 4759 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867978 4759 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.867991 4759 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868003 4759 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868014 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868024 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868034 4759 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868045 4759 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868056 4759 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868067 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868077 4759 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868089 4759 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868102 4759 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868112 4759 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868122 4759 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868132 4759 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868143 4759 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868153 4759 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868167 4759 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868180 4759 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868192 4759 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868203 4759 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868213 4759 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868224 4759 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868235 4759 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868246 4759 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868256 4759 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868266 4759 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868276 4759 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868286 4759 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868297 4759 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868307 4759 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 19:33:29 crc kubenswrapper[4759]: W1125 19:33:29.868320 4759 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.868336 4759 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.868728 4759 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.875729 4759 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.875880 4759 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.879096 4759 server.go:997] "Starting client certificate rotation" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.879144 4759 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.879398 4759 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 12:53:16.739014453 +0000 UTC Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.879578 4759 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 545h19m46.859442884s for next certificate rotation Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.901641 4759 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.907951 4759 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.932522 4759 log.go:25] "Validated CRI v1 runtime API" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.970877 4759 log.go:25] "Validated CRI v1 image API" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.973102 4759 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.981438 4759 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-19-28-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 19:33:29 crc kubenswrapper[4759]: I1125 19:33:29.981527 4759 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.014517 4759 manager.go:217] Machine: {Timestamp:2025-11-25 19:33:30.008272528 +0000 UTC m=+0.594220309 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269 BootID:fb282a89-ea25-4f14-b040-5cf9a95acf23 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:12:3d:c5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:12:3d:c5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:38:5e:86 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9c:41:76 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:91:a9:75 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:19:e6:81 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:aa:7c:50:a8:38:d4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f2:64:1f:ac:44:b6 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.014972 4759 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.015286 4759 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.015947 4759 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.016309 4759 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.016387 4759 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.016928 4759 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.016954 4759 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.017642 4759 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.017711 4759 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.018656 4759 state_mem.go:36] "Initialized new in-memory state store" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.018829 4759 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.022369 4759 kubelet.go:418] "Attempting to sync node with API server" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.022404 4759 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.022466 4759 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.022488 4759 kubelet.go:324] "Adding apiserver pod source" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.022506 4759 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.027199 4759 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.028149 4759 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.029184 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.029330 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.029759 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.029925 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.030177 4759 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031707 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031735 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031746 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031755 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031772 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031781 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031789 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031805 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031814 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031824 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031837 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.031845 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.033490 4759 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.034199 4759 server.go:1280] "Started kubelet" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.034875 4759 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:30 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.036860 4759 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.036860 4759 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.037523 4759 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038614 4759 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038656 4759 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038789 4759 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 12:51:49.784490442 +0000 UTC Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038853 4759 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038863 4759 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038959 4759 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.038853 4759 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 641h18m19.745642002s for next certificate rotation Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.039749 4759 server.go:460] "Adding debug handlers to kubelet server" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.040711 4759 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.041642 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="200ms" Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.041996 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.042072 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.046893 4759 factory.go:55] Registering systemd factory Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.046934 4759 factory.go:221] Registration of the systemd container factory successfully Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.043807 4759 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.41:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b56e28aa2550e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 19:33:30.034173198 +0000 UTC m=+0.620120909,LastTimestamp:2025-11-25 19:33:30.034173198 +0000 UTC m=+0.620120909,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.048466 4759 factory.go:153] Registering CRI-O factory Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.048506 4759 factory.go:221] Registration of the crio container factory successfully Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.048592 4759 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.048804 4759 factory.go:103] Registering Raw factory Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.048825 4759 manager.go:1196] Started watching for new ooms in manager Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.049542 4759 manager.go:319] Starting recovery of all containers Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051666 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051710 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051720 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051731 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051740 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051750 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051773 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051783 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051793 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051802 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051811 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051821 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051835 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051847 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051858 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051867 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051876 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051885 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051893 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051920 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051958 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051967 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051976 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051985 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.051994 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052004 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052034 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052045 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052053 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052062 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052074 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052089 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052098 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052107 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052116 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052125 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052135 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052144 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052156 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052169 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052179 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052189 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052199 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052209 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052218 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052247 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052259 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052270 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052280 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052290 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052301 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052312 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052326 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052338 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052348 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052358 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052369 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052379 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052405 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052414 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052424 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052457 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052469 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052478 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052487 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052496 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052507 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.052518 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054566 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054593 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054604 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054614 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054624 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054633 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054645 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054655 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054664 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054674 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054685 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.054696 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055174 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055196 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055209 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055220 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055232 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055241 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055252 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055261 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.055272 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058187 4759 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058828 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058867 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058894 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058925 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058958 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.058984 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059098 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059201 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059232 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059257 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059315 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059392 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059475 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059508 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059537 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059584 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059616 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059649 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059678 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059704 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059773 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059805 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059845 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059880 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059911 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059937 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059965 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.059991 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060016 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060042 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060066 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060133 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060169 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060193 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060219 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060245 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060281 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060314 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060341 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060383 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060426 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060522 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060550 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060577 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060601 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060648 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060686 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060714 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060762 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060796 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060831 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060877 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060919 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060946 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.060970 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061007 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061034 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061057 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061104 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061129 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061154 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061179 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061218 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061241 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061268 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061292 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061317 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061343 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061370 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061405 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061429 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061489 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061519 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061555 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061580 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061604 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061629 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061654 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061708 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061731 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061770 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061795 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061821 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061846 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061873 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061896 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.061923 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062019 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062058 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062085 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062111 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062136 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062163 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062190 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062213 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062241 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062267 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062291 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062333 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062361 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062389 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062414 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062439 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062497 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062531 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062560 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062584 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062612 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062637 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062663 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062689 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062714 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062740 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062766 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062793 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062820 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062859 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062886 4759 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062910 4759 reconstruct.go:97] "Volume reconstruction finished" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.062927 4759 reconciler.go:26] "Reconciler: start to sync state" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.071208 4759 manager.go:324] Recovery completed Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.083925 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.086273 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.086320 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.086333 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.087009 4759 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.087028 4759 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.087054 4759 state_mem.go:36] "Initialized new in-memory state store" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.101867 4759 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.103151 4759 policy_none.go:49] "None policy: Start" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.105092 4759 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.105141 4759 state_mem.go:35] "Initializing new in-memory state store" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.105553 4759 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.105610 4759 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.105643 4759 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.105704 4759 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.109624 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.109733 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.141039 4759 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.180659 4759 manager.go:334] "Starting Device Plugin manager" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.180722 4759 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.180739 4759 server.go:79] "Starting device plugin registration server" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.181222 4759 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.181243 4759 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.181802 4759 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.181879 4759 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.181885 4759 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.190258 4759 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.206679 4759 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.206802 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208067 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208129 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208357 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208539 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.208619 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209440 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209559 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209572 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209786 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209941 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209973 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.209989 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.210128 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.210232 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211042 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211075 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211087 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211229 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211349 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211387 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211764 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211795 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.211805 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212184 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212215 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212227 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212473 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212505 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212514 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212661 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212792 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.212834 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213708 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213742 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213750 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213729 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213972 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.213999 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.214834 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.214857 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.214865 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.243618 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="400ms" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265082 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265115 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265162 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265193 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265237 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265253 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265269 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265291 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265372 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265424 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265481 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265527 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265573 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265596 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.265629 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.282291 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.284017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.284093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.284113 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.284145 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.284714 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.41:6443: connect: connection refused" node="crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367172 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367283 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367330 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367373 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367417 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367513 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367425 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367520 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367514 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367518 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367612 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367658 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367695 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367708 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367729 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367765 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367754 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367799 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367801 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367780 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367855 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367836 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367879 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367836 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.367956 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.368007 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.368018 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.368042 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.368074 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.368154 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.485414 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.487147 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.487216 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.487238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.487279 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.488107 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.41:6443: connect: connection refused" node="crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.548027 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.562540 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.584891 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.594274 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.600820 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.609207 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-fe5581625b7a85c8e956614834207c4cd3c6f7d4d7b96e13ce9117c598398676 WatchSource:0}: Error finding container fe5581625b7a85c8e956614834207c4cd3c6f7d4d7b96e13ce9117c598398676: Status 404 returned error can't find the container with id fe5581625b7a85c8e956614834207c4cd3c6f7d4d7b96e13ce9117c598398676 Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.626424 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0f0b32920260b4deb065890f8eb0e3742983966d3034dde93fa8ea9228e24f21 WatchSource:0}: Error finding container 0f0b32920260b4deb065890f8eb0e3742983966d3034dde93fa8ea9228e24f21: Status 404 returned error can't find the container with id 0f0b32920260b4deb065890f8eb0e3742983966d3034dde93fa8ea9228e24f21 Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.632877 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e253c3a4278523399c86d7169e5ad99b2c05d5e03a44b0bca31fca08e6c8bff1 WatchSource:0}: Error finding container e253c3a4278523399c86d7169e5ad99b2c05d5e03a44b0bca31fca08e6c8bff1: Status 404 returned error can't find the container with id e253c3a4278523399c86d7169e5ad99b2c05d5e03a44b0bca31fca08e6c8bff1 Nov 25 19:33:30 crc kubenswrapper[4759]: W1125 19:33:30.634277 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ed2c06004c25f7631c8c50d075d368314b2096ee43252e61c293c582b48bf8fe WatchSource:0}: Error finding container ed2c06004c25f7631c8c50d075d368314b2096ee43252e61c293c582b48bf8fe: Status 404 returned error can't find the container with id ed2c06004c25f7631c8c50d075d368314b2096ee43252e61c293c582b48bf8fe Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.644536 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="800ms" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.888532 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.890272 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.890344 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.890364 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:30 crc kubenswrapper[4759]: I1125 19:33:30.890408 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:30 crc kubenswrapper[4759]: E1125 19:33:30.891133 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.41:6443: connect: connection refused" node="crc" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.036721 4759 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:31 crc kubenswrapper[4759]: W1125 19:33:31.069605 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.069688 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.110148 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e253c3a4278523399c86d7169e5ad99b2c05d5e03a44b0bca31fca08e6c8bff1"} Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.111669 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f0b32920260b4deb065890f8eb0e3742983966d3034dde93fa8ea9228e24f21"} Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.112838 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe5581625b7a85c8e956614834207c4cd3c6f7d4d7b96e13ce9117c598398676"} Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.114133 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3f1345b41248ed6c07bcf91a399bb6ea962265bcb0271b667308b20661874b06"} Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.115212 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ed2c06004c25f7631c8c50d075d368314b2096ee43252e61c293c582b48bf8fe"} Nov 25 19:33:31 crc kubenswrapper[4759]: W1125 19:33:31.222992 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.223413 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:31 crc kubenswrapper[4759]: W1125 19:33:31.321195 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.321303 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.445334 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="1.6s" Nov 25 19:33:31 crc kubenswrapper[4759]: W1125 19:33:31.453514 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.453662 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.691418 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.694205 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.694255 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.694273 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:31 crc kubenswrapper[4759]: I1125 19:33:31.694308 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:31 crc kubenswrapper[4759]: E1125 19:33:31.695002 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.41:6443: connect: connection refused" node="crc" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.035844 4759 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.119890 4759 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd" exitCode=0 Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.119980 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.119996 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.121399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.121470 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.121489 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.122091 4759 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef" exitCode=0 Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.122187 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.122195 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.123316 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.123362 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.123375 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.127267 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.127245 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.127507 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.127536 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.127589 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.128158 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.128201 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.128219 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.128926 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f" exitCode=0 Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.129005 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.129027 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.129863 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.129927 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.129945 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.131011 4759 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5" exitCode=0 Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.131053 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5"} Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.131126 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.132178 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.132223 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.132237 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.136890 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.138298 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.138323 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:32 crc kubenswrapper[4759]: I1125 19:33:32.138334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:32 crc kubenswrapper[4759]: E1125 19:33:32.346861 4759 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.41:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b56e28aa2550e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 19:33:30.034173198 +0000 UTC m=+0.620120909,LastTimestamp:2025-11-25 19:33:30.034173198 +0000 UTC m=+0.620120909,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.036854 4759 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:33 crc kubenswrapper[4759]: E1125 19:33:33.046724 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="3.2s" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.137395 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.137508 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.139537 4759 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc" exitCode=0 Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.139646 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.139757 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.141089 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.141148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.141175 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.144062 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"04d5d6e432bf167987c3da881c52aecd060161d9a794e0545a58bf2a31ee9ba8"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.144094 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.145331 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.145380 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.145398 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.147377 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.147375 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.147553 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82"} Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.148360 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.148427 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.148496 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.256935 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.295331 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.297023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.297065 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.297077 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:33 crc kubenswrapper[4759]: I1125 19:33:33.297105 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:33 crc kubenswrapper[4759]: E1125 19:33:33.297765 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.41:6443: connect: connection refused" node="crc" Nov 25 19:33:33 crc kubenswrapper[4759]: W1125 19:33:33.322270 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:33 crc kubenswrapper[4759]: E1125 19:33:33.322357 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:33 crc kubenswrapper[4759]: W1125 19:33:33.748785 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:33 crc kubenswrapper[4759]: E1125 19:33:33.748851 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:33 crc kubenswrapper[4759]: W1125 19:33:33.927965 4759 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:33 crc kubenswrapper[4759]: E1125 19:33:33.928083 4759 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.41:6443: connect: connection refused" logger="UnhandledError" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.036556 4759 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.41:6443: connect: connection refused Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.154039 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb"} Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.154092 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838"} Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.154104 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9"} Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.154215 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.155649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.155677 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.155686 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.156496 4759 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4" exitCode=0 Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.156551 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4"} Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.156703 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.157988 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.158008 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.158017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.161175 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.161637 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.161884 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425"} Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.161945 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162529 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162549 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162557 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162962 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.162969 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.163238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.163260 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:34 crc kubenswrapper[4759]: I1125 19:33:34.163269 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169838 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"85f46c4fd1e2e9f85df12f3b4e03d903a1f2eae9ba61950682af19c45d0964aa"} Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169893 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169907 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"df09a503ad4d85bb0383d070f52ac979188cbd361d3dc318ae1e96742daebf4a"} Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169932 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa4aa15e3a239ee31d509b20eac95c20c7a636b6c7c2c80cf65f5b386ca07d64"} Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169933 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.170011 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.169950 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171580 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171617 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171634 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171639 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171689 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.171706 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.559144 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.559345 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.560785 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.560860 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.560884 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.564942 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:35 crc kubenswrapper[4759]: I1125 19:33:35.817255 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.178609 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.178609 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.178597 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f7f20ac2001ecb232136b91658f51968bee70d612dd01f7be1708b228a686e0d"} Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.179417 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ee63675d5b4dac84d9cc76b94052791eae19e67b1b6ca63b5641bb45731df462"} Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180059 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180090 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180138 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180139 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180168 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.180186 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.257576 4759 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.257696 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.274970 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.275232 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.275322 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.279717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.279801 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.279825 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.498754 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.500662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.500723 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.500750 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.500793 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:36 crc kubenswrapper[4759]: I1125 19:33:36.977347 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.181402 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.181433 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.181530 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.181611 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183541 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183572 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183597 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183600 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183548 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183710 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183728 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:37 crc kubenswrapper[4759]: I1125 19:33:37.183609 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:39 crc kubenswrapper[4759]: I1125 19:33:39.933556 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 19:33:39 crc kubenswrapper[4759]: I1125 19:33:39.933782 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:39 crc kubenswrapper[4759]: I1125 19:33:39.935234 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:39 crc kubenswrapper[4759]: I1125 19:33:39.935291 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:39 crc kubenswrapper[4759]: I1125 19:33:39.935319 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:40 crc kubenswrapper[4759]: E1125 19:33:40.190371 4759 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.211385 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.211560 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.212635 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.212704 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.212724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.328775 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.328984 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.330288 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.330354 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.330377 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.442922 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.443156 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.444528 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.444599 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:40 crc kubenswrapper[4759]: I1125 19:33:40.444620 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.496201 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.496724 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.502636 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.502687 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.782678 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.782911 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.783929 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.783968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.783979 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.817717 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 19:33:44 crc kubenswrapper[4759]: I1125 19:33:44.946539 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.202049 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.203394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.203439 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.203471 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.823010 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.823135 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.824406 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.824483 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:45 crc kubenswrapper[4759]: I1125 19:33:45.824494 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.204865 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.206576 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.206676 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.206695 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.257766 4759 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.257855 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.281719 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.281930 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.283239 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.283288 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.283304 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:46 crc kubenswrapper[4759]: I1125 19:33:46.286937 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:47 crc kubenswrapper[4759]: I1125 19:33:47.207853 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:47 crc kubenswrapper[4759]: I1125 19:33:47.209125 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:47 crc kubenswrapper[4759]: I1125 19:33:47.209244 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:47 crc kubenswrapper[4759]: I1125 19:33:47.209282 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:49 crc kubenswrapper[4759]: E1125 19:33:49.487042 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.490544 4759 trace.go:236] Trace[7421438]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 19:33:34.514) (total time: 14975ms): Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[7421438]: ---"Objects listed" error: 14975ms (19:33:49.490) Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[7421438]: [14.975750243s] [14.975750243s] END Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.490593 4759 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.491889 4759 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.491959 4759 trace.go:236] Trace[75860926]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 19:33:37.018) (total time: 12473ms): Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[75860926]: ---"Objects listed" error: 12473ms (19:33:49.491) Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[75860926]: [12.473172294s] [12.473172294s] END Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.491988 4759 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.492375 4759 trace.go:236] Trace[934912624]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 19:33:37.363) (total time: 12128ms): Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[934912624]: ---"Objects listed" error: 12128ms (19:33:49.492) Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[934912624]: [12.128947878s] [12.128947878s] END Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.492403 4759 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.494423 4759 trace.go:236] Trace[278789085]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 19:33:37.847) (total time: 11646ms): Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[278789085]: ---"Objects listed" error: 11646ms (19:33:49.494) Nov 25 19:33:49 crc kubenswrapper[4759]: Trace[278789085]: [11.646509918s] [11.646509918s] END Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.494486 4759 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 19:33:49 crc kubenswrapper[4759]: E1125 19:33:49.494556 4759 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539115 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58868->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539187 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58868->192.168.126.11:17697: read: connection reset by peer" Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539694 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58876->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539720 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539769 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58876->192.168.126.11:17697: read: connection reset by peer" Nov 25 19:33:49 crc kubenswrapper[4759]: I1125 19:33:49.539804 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.034634 4759 apiserver.go:52] "Watching apiserver" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.039250 4759 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.039806 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.040344 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.040554 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.040689 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.040816 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.040907 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.040985 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.041049 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.041197 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.041331 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.043525 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046138 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046187 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046256 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046307 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046662 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046724 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.046911 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.051172 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.078934 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.102856 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.127967 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.140651 4759 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.146799 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.158286 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.170994 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.181172 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.196004 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197228 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197414 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197564 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197751 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197886 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.197984 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198075 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198185 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198303 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198414 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198543 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198645 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198775 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198890 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199025 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199135 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198114 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198593 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199295 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199322 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199338 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199251 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199349 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198771 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198909 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199415 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199515 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199553 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199576 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199597 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199615 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199635 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199652 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.198982 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199013 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199163 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199711 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199602 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199639 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199670 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199788 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199812 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199836 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199837 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199868 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199873 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.199891 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200013 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200031 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200037 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200057 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200071 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200124 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200122 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200227 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200168 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200325 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200370 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200407 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200433 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200487 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200520 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200546 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200610 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200642 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200372 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200425 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200625 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200649 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200772 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200787 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200900 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200907 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200670 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200929 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200959 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.200990 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201018 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201060 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201087 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201112 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201138 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201163 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201187 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201214 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201238 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201262 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201318 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201345 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201371 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201397 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201479 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201509 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201536 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201561 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201586 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201641 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201666 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201694 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201717 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201742 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201768 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201795 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201820 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201851 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201880 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201906 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201930 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201955 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201975 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201999 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202022 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202044 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202069 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202089 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202110 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202131 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202153 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202197 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202219 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202241 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202261 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202281 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202305 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202328 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202370 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202394 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204907 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204942 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201108 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204969 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201122 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201183 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201318 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204995 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205000 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205219 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205253 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205278 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205300 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205323 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205345 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205365 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205385 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205407 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205428 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205466 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205485 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205504 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205527 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205548 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205567 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205585 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205620 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205647 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205667 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205687 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205705 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205722 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205741 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205760 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205778 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205796 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205222 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201429 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201585 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201599 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201703 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201732 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201903 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.201976 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202333 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.202748 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.203204 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.203233 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.203372 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.203817 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.203862 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204348 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204607 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.204958 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205169 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205299 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205342 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205400 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205461 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205482 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205634 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205676 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205751 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205790 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.205816 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206093 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206122 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206160 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206184 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206209 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206236 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206262 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206289 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206320 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206360 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206394 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206416 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206462 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206502 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206526 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206549 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206573 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206597 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206620 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206642 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206664 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206691 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206731 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206765 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206789 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206812 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206838 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206862 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206884 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206907 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206933 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206955 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206976 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207000 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207022 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207046 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207069 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207092 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207117 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207141 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207165 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207189 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207213 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207236 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207259 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207287 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207312 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207336 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207359 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207384 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207406 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207440 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207522 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207564 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207595 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207687 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207905 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207939 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207971 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208005 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208040 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208073 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208109 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208142 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208168 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208191 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208240 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208274 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208303 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208327 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208366 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208392 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208415 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208474 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208503 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208528 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208555 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209340 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209382 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209408 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209501 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209527 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209547 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209566 4759 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209588 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209644 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209662 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209684 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209706 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209725 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209745 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209759 4759 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209772 4759 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209786 4759 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209799 4759 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209813 4759 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209826 4759 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209846 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209865 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209883 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209902 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209947 4759 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209962 4759 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209979 4759 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.209999 4759 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210017 4759 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210036 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210055 4759 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210073 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210092 4759 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210111 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210130 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210149 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210162 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210175 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210189 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210203 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210217 4759 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210240 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210254 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210268 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210282 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210298 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210313 4759 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210327 4759 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210339 4759 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211362 4759 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206212 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206304 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.206733 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207178 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207529 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.207536 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.208585 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210277 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210213 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210403 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210417 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210540 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210597 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210712 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210890 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.210919 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211068 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211489 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211570 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211712 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211728 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211809 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.211831 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.212014 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.212112 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.213351 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.213513 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.213529 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.213865 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.213910 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214044 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214133 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214292 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214405 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214491 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214493 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214711 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.214857 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215005 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215274 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215373 4759 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.219084 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.222513 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.222541 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215464 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.222654 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215927 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215959 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215973 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.216126 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.216584 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.216731 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.218117 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.222735 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.218228 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.222940 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.223178 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.223466 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.215347 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.223704 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.223724 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.223900 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.224103 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.224374 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.225753 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.225933 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.226058 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.230134 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.230188 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.230318 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.230585 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.230635 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.231068 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.231309 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.231584 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.232468 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.232612 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.232699 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.233357 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.233362 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.233582 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:33:50.733528736 +0000 UTC m=+21.319476527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.233642 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.233739 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.233808 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:50.733788553 +0000 UTC m=+21.319736444 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.233834 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.234315 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.234731 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.235095 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.235168 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:50.735154172 +0000 UTC m=+21.321101883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.236202 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.236529 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.236967 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.237392 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.237959 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.238146 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.238422 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.238529 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.239081 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.239109 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.239205 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.239496 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.239697 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.240747 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.244567 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.245193 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.245571 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb" exitCode=255 Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.245637 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb"} Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.245944 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.246160 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.246175 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.246188 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.246178 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.246236 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:50.746221018 +0000 UTC m=+21.332168719 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.246897 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.247302 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.247560 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.249223 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.250787 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.250808 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.250817 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.250863 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:50.750840425 +0000 UTC m=+21.336788126 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.253987 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.255094 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.255978 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.256152 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.256275 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.256336 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.257562 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.257811 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.259794 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.259866 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.261953 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.262551 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.263631 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.263758 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.263786 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.263787 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.264245 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.264480 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.264655 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.265222 4759 scope.go:117] "RemoveContainer" containerID="b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.266972 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.267008 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.267069 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.267115 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.267307 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.269293 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.269994 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.270482 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.270851 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.271476 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.272473 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.273195 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.275831 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.276110 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.276590 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.277137 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.280596 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.281322 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.286036 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.290062 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.291485 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.293052 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.298504 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.299298 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.309481 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311012 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311055 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311110 4759 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311121 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311133 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311142 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311151 4759 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311161 4759 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311169 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311177 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311186 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311195 4759 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311204 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311214 4759 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311222 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311232 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311241 4759 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311249 4759 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311257 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311266 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311275 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311286 4759 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311295 4759 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311303 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311302 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311312 4759 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311367 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311379 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311390 4759 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311407 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311418 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311428 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311438 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311462 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311471 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311480 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311489 4759 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311498 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311264 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311507 4759 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311550 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311563 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311573 4759 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311584 4759 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311593 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311602 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311612 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311623 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311633 4759 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311642 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311651 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311660 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311672 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311683 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311692 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311701 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311710 4759 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311720 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311729 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311738 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311747 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311759 4759 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311769 4759 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311778 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311788 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311801 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311812 4759 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311820 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311829 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311838 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311849 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311859 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311868 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311877 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311887 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311895 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311904 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311913 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311923 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311934 4759 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311943 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311952 4759 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311961 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311969 4759 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311978 4759 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311987 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.311996 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312007 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312017 4759 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312025 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312035 4759 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312044 4759 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312052 4759 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312061 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312071 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312079 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312088 4759 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312096 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312106 4759 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312115 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312123 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312132 4759 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312141 4759 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312151 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312163 4759 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312172 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312180 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312189 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312198 4759 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312207 4759 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312217 4759 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312225 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312234 4759 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312242 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312251 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312260 4759 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312270 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312279 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312287 4759 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312296 4759 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312305 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312313 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312321 4759 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312330 4759 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312338 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312347 4759 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312356 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312597 4759 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312609 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312618 4759 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312626 4759 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312635 4759 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312644 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312653 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312662 4759 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312671 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312680 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312689 4759 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312698 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312707 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312716 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312725 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312733 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312744 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312753 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312762 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312771 4759 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312780 4759 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312789 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312797 4759 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312805 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.312814 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.317635 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.327157 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.336798 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.346480 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.356131 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.356534 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.364312 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 19:33:50 crc kubenswrapper[4759]: W1125 19:33:50.368741 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-8feea05eb575a7d1cc85c4975f71d14f58ab4a7858d71e59b268c18b9d64d2ee WatchSource:0}: Error finding container 8feea05eb575a7d1cc85c4975f71d14f58ab4a7858d71e59b268c18b9d64d2ee: Status 404 returned error can't find the container with id 8feea05eb575a7d1cc85c4975f71d14f58ab4a7858d71e59b268c18b9d64d2ee Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.372650 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.716851 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-jz4pb"] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.717508 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ttvrw"] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.717676 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.717747 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.719830 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.719840 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.720231 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.720464 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.720619 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.720719 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.721067 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.731345 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.745751 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.757106 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.766019 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.782993 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.806003 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817058 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817177 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfmwd\" (UniqueName: \"kubernetes.io/projected/bb894524-1c88-4838-aa0c-0a1956951d23-kube-api-access-bfmwd\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817220 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817278 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817309 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb894524-1c88-4838-aa0c-0a1956951d23-hosts-file\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817345 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3822015e-9df0-4cce-9d59-92247b44348b-serviceca\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817373 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qp56\" (UniqueName: \"kubernetes.io/projected/3822015e-9df0-4cce-9d59-92247b44348b-kube-api-access-5qp56\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817402 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817421 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3822015e-9df0-4cce-9d59-92247b44348b-host\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.817518 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827190 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827327 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827395 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:51.827367134 +0000 UTC m=+22.413314835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827461 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:51.827428145 +0000 UTC m=+22.413375846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827482 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827518 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827535 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827604 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.827614 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rfsr5"] Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827645 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827664 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827610 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:33:51.82758015 +0000 UTC m=+22.413527851 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827708 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:51.827686463 +0000 UTC m=+22.413634164 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: E1125 19:33:50.827745 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:51.827732994 +0000 UTC m=+22.413680705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.828570 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.834243 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.834274 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vz6pj"] Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.834482 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.834516 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.835279 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.839910 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.843340 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.843476 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.843583 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.843913 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.844070 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.844844 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.852760 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.861828 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.873896 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.884169 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.898947 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.906869 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.917812 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918148 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-k8s-cni-cncf-io\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918283 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qp56\" (UniqueName: \"kubernetes.io/projected/3822015e-9df0-4cce-9d59-92247b44348b-kube-api-access-5qp56\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918415 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-daemon-config\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918537 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-bin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918636 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/07c1b558-ca9e-4edc-9420-05750ceee3a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918725 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-system-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918827 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-multus\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.918932 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-kubelet\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919044 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/07c1b558-ca9e-4edc-9420-05750ceee3a9-rootfs\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919147 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-netns\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919367 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-os-release\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919502 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cni-binary-copy\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919540 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-conf-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919579 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3822015e-9df0-4cce-9d59-92247b44348b-host\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919604 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-hostroot\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919639 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-etc-kubernetes\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919664 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8q88\" (UniqueName: \"kubernetes.io/projected/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-kube-api-access-s8q88\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919689 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdj5f\" (UniqueName: \"kubernetes.io/projected/07c1b558-ca9e-4edc-9420-05750ceee3a9-kube-api-access-sdj5f\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919718 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfmwd\" (UniqueName: \"kubernetes.io/projected/bb894524-1c88-4838-aa0c-0a1956951d23-kube-api-access-bfmwd\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919743 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919766 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-socket-dir-parent\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919789 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-multus-certs\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919814 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/07c1b558-ca9e-4edc-9420-05750ceee3a9-proxy-tls\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919859 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb894524-1c88-4838-aa0c-0a1956951d23-hosts-file\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919883 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3822015e-9df0-4cce-9d59-92247b44348b-serviceca\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919904 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cnibin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.919984 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3822015e-9df0-4cce-9d59-92247b44348b-host\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.920197 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb894524-1c88-4838-aa0c-0a1956951d23-hosts-file\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.921063 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3822015e-9df0-4cce-9d59-92247b44348b-serviceca\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.935299 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.966868 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.977943 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.980682 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qp56\" (UniqueName: \"kubernetes.io/projected/3822015e-9df0-4cce-9d59-92247b44348b-kube-api-access-5qp56\") pod \"node-ca-ttvrw\" (UID: \"3822015e-9df0-4cce-9d59-92247b44348b\") " pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.980791 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfmwd\" (UniqueName: \"kubernetes.io/projected/bb894524-1c88-4838-aa0c-0a1956951d23-kube-api-access-bfmwd\") pod \"node-resolver-jz4pb\" (UID: \"bb894524-1c88-4838-aa0c-0a1956951d23\") " pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.987885 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:50 crc kubenswrapper[4759]: I1125 19:33:50.996530 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.005247 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.014286 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020644 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020682 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-socket-dir-parent\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020697 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-multus-certs\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020714 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/07c1b558-ca9e-4edc-9420-05750ceee3a9-proxy-tls\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020731 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cnibin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020787 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-k8s-cni-cncf-io\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020806 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-daemon-config\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020824 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-bin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020840 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/07c1b558-ca9e-4edc-9420-05750ceee3a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020861 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-system-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020881 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-multus\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020882 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-multus-certs\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020898 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-kubelet\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020950 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cnibin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020987 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-system-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.020958 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-socket-dir-parent\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021063 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-bin\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021027 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/07c1b558-ca9e-4edc-9420-05750ceee3a9-rootfs\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021026 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-cni-multus\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021001 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/07c1b558-ca9e-4edc-9420-05750ceee3a9-rootfs\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021241 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-netns\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021155 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-cni-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021268 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-netns\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021268 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-os-release\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021332 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cni-binary-copy\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021347 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-os-release\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021359 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-conf-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021391 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-hostroot\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021415 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-etc-kubernetes\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021432 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8q88\" (UniqueName: \"kubernetes.io/projected/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-kube-api-access-s8q88\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021470 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdj5f\" (UniqueName: \"kubernetes.io/projected/07c1b558-ca9e-4edc-9420-05750ceee3a9-kube-api-access-sdj5f\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021536 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-hostroot\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021592 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/07c1b558-ca9e-4edc-9420-05750ceee3a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021611 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-conf-dir\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021646 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-etc-kubernetes\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.021681 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-var-lib-kubelet\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.022137 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-multus-daemon-config\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.022145 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-cni-binary-copy\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.022225 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-host-run-k8s-cni-cncf-io\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.025373 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/07c1b558-ca9e-4edc-9420-05750ceee3a9-proxy-tls\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.031158 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jz4pb" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.036329 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ttvrw" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.037490 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8q88\" (UniqueName: \"kubernetes.io/projected/67e7b861-9f72-41ba-ab98-35e0a2bdbe39-kube-api-access-s8q88\") pod \"multus-vz6pj\" (UID: \"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\") " pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.038928 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdj5f\" (UniqueName: \"kubernetes.io/projected/07c1b558-ca9e-4edc-9420-05750ceee3a9-kube-api-access-sdj5f\") pod \"machine-config-daemon-rfsr5\" (UID: \"07c1b558-ca9e-4edc-9420-05750ceee3a9\") " pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: W1125 19:33:51.044215 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb894524_1c88_4838_aa0c_0a1956951d23.slice/crio-e8a516df07a6690dd15da141671f20527c0610d3d1f8a2276c453904fb57302c WatchSource:0}: Error finding container e8a516df07a6690dd15da141671f20527c0610d3d1f8a2276c453904fb57302c: Status 404 returned error can't find the container with id e8a516df07a6690dd15da141671f20527c0610d3d1f8a2276c453904fb57302c Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.106337 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.106469 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.142297 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.149572 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vz6pj" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.254805 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-58fww"] Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.255385 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.258734 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.259071 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.277790 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.286266 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.297223 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.297406 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.298760 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerStarted","Data":"51907eb69766c3fe9bc99684bdaa4bbbe7bd1304d42789976a34ef77ebe31f9f"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.303738 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ttvrw" event={"ID":"3822015e-9df0-4cce-9d59-92247b44348b","Type":"ContainerStarted","Data":"439cf35bd57adbaab37d91e15b4c58bff615019b512ffd77bfdb0d2d94cefe45"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.305507 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jz4pb" event={"ID":"bb894524-1c88-4838-aa0c-0a1956951d23","Type":"ContainerStarted","Data":"e8a516df07a6690dd15da141671f20527c0610d3d1f8a2276c453904fb57302c"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.309211 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.309268 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.309284 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"04a207d87538581be691b1c62dc2643ed24521adc616f51358df6c9efed08bad"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.310910 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.310947 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8feea05eb575a7d1cc85c4975f71d14f58ab4a7858d71e59b268c18b9d64d2ee"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.313776 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"158f4eff9047badaf51c07bafdac120c84c2ca7f6d187e6c20bcf61b9edaafde"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.314299 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.317772 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"68a81db22cf7d44d995b203aa5b638ec952b5d44ae168d5314e5d665e92d2760"} Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.327598 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.361995 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.379081 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.393706 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.409618 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.426844 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-system-cni-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.426894 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-os-release\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.426941 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.427011 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2ndv\" (UniqueName: \"kubernetes.io/projected/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-kube-api-access-s2ndv\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.427041 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cnibin\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.427074 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.427104 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.430924 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.449946 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.463558 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.477330 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.497357 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.512506 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.524159 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.527620 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-system-cni-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.527659 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-os-release\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.527696 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.527731 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-system-cni-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.527761 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2ndv\" (UniqueName: \"kubernetes.io/projected/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-kube-api-access-s2ndv\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528045 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cnibin\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528048 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-os-release\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528126 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528196 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528190 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cnibin\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528375 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528504 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.528730 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.546944 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2ndv\" (UniqueName: \"kubernetes.io/projected/6dff2992-ff1a-4dce-b635-ee0f16ba11d8-kube-api-access-s2ndv\") pod \"multus-additional-cni-plugins-58fww\" (UID: \"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\") " pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.568028 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.588963 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-58fww" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.599600 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.623870 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.639896 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.660303 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.675065 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-99nmn"] Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.676002 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.678553 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.678885 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.678946 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.679065 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.679246 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.679297 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.679604 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.689903 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.705255 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.721919 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.737809 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.750612 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.769632 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.782208 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.795544 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.813299 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.828938 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831352 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831495 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831527 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.831630 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:33:53.83158769 +0000 UTC m=+24.417535391 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831742 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831813 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831865 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831930 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.831978 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832021 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832046 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832074 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832175 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832201 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832218 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832216 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832256 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832268 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:53.832259439 +0000 UTC m=+24.418207140 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832293 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832322 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lr5j\" (UniqueName: \"kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832343 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832363 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832384 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832426 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832482 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832529 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832590 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832667 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832904 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.832941 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832951 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.832976 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833007 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:53.83299826 +0000 UTC m=+24.418945961 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833041 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:53.83302147 +0000 UTC m=+24.418969361 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833173 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833203 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833218 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:51 crc kubenswrapper[4759]: E1125 19:33:51.833283 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:53.833264977 +0000 UTC m=+24.419212678 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.845652 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.860877 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.876374 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.905858 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934296 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934363 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934396 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934390 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934482 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934510 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934515 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934571 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934570 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934588 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934606 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934618 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934638 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934667 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934681 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934696 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934703 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934712 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934731 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934750 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934753 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934780 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934781 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934803 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934807 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934826 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934861 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934887 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934891 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934729 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.934944 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lr5j\" (UniqueName: \"kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935036 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935116 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935153 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935198 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935264 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935289 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.935880 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.940258 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.945853 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.973352 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lr5j\" (UniqueName: \"kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j\") pod \"ovnkube-node-99nmn\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:51 crc kubenswrapper[4759]: I1125 19:33:51.991641 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:52 crc kubenswrapper[4759]: W1125 19:33:52.001499 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8b6037c_81e1_4fe8_9216_5b1863a029f1.slice/crio-cd42469ea6973db6d014ce9669d6690b8067361de911ac8e65bdab3b5cd8fe84 WatchSource:0}: Error finding container cd42469ea6973db6d014ce9669d6690b8067361de911ac8e65bdab3b5cd8fe84: Status 404 returned error can't find the container with id cd42469ea6973db6d014ce9669d6690b8067361de911ac8e65bdab3b5cd8fe84 Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.015751 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.046914 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.085883 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.106321 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.106393 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:52 crc kubenswrapper[4759]: E1125 19:33:52.106552 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:33:52 crc kubenswrapper[4759]: E1125 19:33:52.106725 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.110390 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.111154 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.112297 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.113001 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.114016 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.114699 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.115409 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.116605 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.117464 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.118648 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.119247 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.120629 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.121232 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.121884 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.123101 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.123738 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.124941 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.125462 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.126180 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.127429 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.128104 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.129350 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.129852 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.131052 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.131562 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.132249 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.133591 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.134143 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.135337 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.135992 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.137065 4759 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.137324 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.138950 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.139775 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.140250 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.142865 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.143582 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.144147 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.145379 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.146962 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.147932 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.148617 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.149316 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.150013 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.150556 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.151151 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.151773 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.152552 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.153083 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.154981 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.155684 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.156586 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.157250 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.157895 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.321115 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerStarted","Data":"105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.322589 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jz4pb" event={"ID":"bb894524-1c88-4838-aa0c-0a1956951d23","Type":"ContainerStarted","Data":"b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.323743 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" exitCode=0 Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.323770 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.323806 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"cd42469ea6973db6d014ce9669d6690b8067361de911ac8e65bdab3b5cd8fe84"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.325374 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182" exitCode=0 Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.325422 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.325455 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerStarted","Data":"e1c7ab713353fe115a1ec2a3a4c4d059e6d39c501919018ab8a8248df3642fe0"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.326866 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ttvrw" event={"ID":"3822015e-9df0-4cce-9d59-92247b44348b","Type":"ContainerStarted","Data":"2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.328553 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.328582 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf"} Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.340133 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.374887 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.397090 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.413834 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.428697 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.446043 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.478929 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.492999 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.510211 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.532380 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.546282 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.564006 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.614304 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.644214 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.686761 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.728541 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.776895 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.808544 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.844255 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.885394 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.925589 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:52 crc kubenswrapper[4759]: I1125 19:33:52.965231 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.009578 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.049191 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.084238 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.105895 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.106005 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.124926 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.261472 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.267698 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.270418 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.276236 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.289743 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.302951 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.318290 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.332558 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.334304 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerStarted","Data":"d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338460 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338484 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338494 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338501 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338510 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.338523 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.354769 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.385397 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.423653 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.466160 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.504260 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.544677 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.585943 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.624123 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.668585 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.703161 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.744709 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.787726 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.823645 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.853020 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.853114 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.853139 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.853163 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.853179 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853233 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:33:57.853195988 +0000 UTC m=+28.439143689 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853271 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853293 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853313 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:57.853300731 +0000 UTC m=+28.439248422 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853319 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853340 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853348 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853357 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853362 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853372 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853346 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:57.853334892 +0000 UTC m=+28.439282593 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853413 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:57.853396444 +0000 UTC m=+28.439344145 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:53 crc kubenswrapper[4759]: E1125 19:33:53.853433 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:33:57.853427014 +0000 UTC m=+28.439374715 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.865989 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.906683 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.950906 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:53 crc kubenswrapper[4759]: I1125 19:33:53.996094 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:53Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.025256 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.068238 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.103952 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.106150 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.106150 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:54 crc kubenswrapper[4759]: E1125 19:33:54.106365 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:33:54 crc kubenswrapper[4759]: E1125 19:33:54.106268 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.144210 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.182797 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.223511 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.345603 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426" exitCode=0 Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.345675 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426"} Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.362758 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.380131 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.405610 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.422778 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.434610 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.465147 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.510760 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.545573 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.588052 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.632131 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.665412 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.708993 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.744434 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:54 crc kubenswrapper[4759]: I1125 19:33:54.791885 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:54Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.106183 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.106397 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.352123 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22" exitCode=0 Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.352185 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22"} Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.365607 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.382245 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.398490 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.410923 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.428208 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.442818 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.456308 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.469710 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.479186 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.495094 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.509651 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.521714 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.534817 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.557342 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.895287 4759 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.896989 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.897034 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.897045 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.897187 4759 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.904715 4759 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.904984 4759 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.905921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.905957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.905965 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.905980 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.905991 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.917165 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.920740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.920775 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.920787 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.920804 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.920817 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.933149 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.935844 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.935865 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.935873 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.935886 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.935895 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.947523 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.950172 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.950206 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.950214 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.950229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.950239 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.965646 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.968681 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.968714 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.968724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.968740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.968749 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.978676 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:55Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:55 crc kubenswrapper[4759]: E1125 19:33:55.978833 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.980238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.980272 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.980282 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.980296 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:55 crc kubenswrapper[4759]: I1125 19:33:55.980306 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:55Z","lastTransitionTime":"2025-11-25T19:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.108467 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.108568 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:56 crc kubenswrapper[4759]: E1125 19:33:56.108622 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:33:56 crc kubenswrapper[4759]: E1125 19:33:56.108713 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.110384 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.110412 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.110421 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.110433 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.110455 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.212345 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.212380 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.212389 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.212404 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.212412 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.314860 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.314909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.314927 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.314951 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.314967 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.358674 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.360909 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6" exitCode=0 Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.360963 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.374474 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.384502 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.395826 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.412105 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.416700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.416727 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.416739 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.416752 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.416760 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.420957 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.430400 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.439419 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.449990 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.462292 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.472485 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.483726 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.496806 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.506157 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519247 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519286 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519295 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519312 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519321 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.519808 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:56Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.621965 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.622002 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.622010 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.622025 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.622036 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.723505 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.723545 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.723557 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.723572 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.723582 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.826208 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.826246 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.826256 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.826270 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.826281 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.929140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.929177 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.929186 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.929200 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:56 crc kubenswrapper[4759]: I1125 19:33:56.929209 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:56Z","lastTransitionTime":"2025-11-25T19:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.031910 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.031951 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.031967 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.031987 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.032001 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.105899 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.106096 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.134037 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.134107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.134130 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.134160 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.134185 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.237403 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.237437 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.237462 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.237477 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.237487 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.340398 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.340431 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.340442 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.340488 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.340501 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.370931 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12" exitCode=0 Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.370977 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.392294 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.410184 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.425821 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.443984 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.444049 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.444074 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.444109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.444131 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.445718 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.465911 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.480778 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547153 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547203 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547217 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547234 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547247 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.547920 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.560239 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.583053 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.597812 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.615649 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.631035 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.643226 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.649916 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.649950 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.649959 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.649974 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.649983 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.655602 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:57Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.752045 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.752078 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.752089 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.752105 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.752119 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.855534 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.855564 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.855573 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.855587 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.855595 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.925956 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.926039 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.926062 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.926084 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.926100 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926205 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:34:05.926179918 +0000 UTC m=+36.512127639 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926202 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926242 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926263 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926276 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926242 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926332 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926340 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926277 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:05.926268131 +0000 UTC m=+36.512215952 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926405 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:05.926383684 +0000 UTC m=+36.512331425 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926430 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926467 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:05.926428135 +0000 UTC m=+36.512375856 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:33:57 crc kubenswrapper[4759]: E1125 19:33:57.926505 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:05.926489837 +0000 UTC m=+36.512437548 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.957693 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.957730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.957741 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.957757 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:57 crc kubenswrapper[4759]: I1125 19:33:57.957774 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:57Z","lastTransitionTime":"2025-11-25T19:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.062051 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.062120 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.062142 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.062171 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.062193 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.106760 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:33:58 crc kubenswrapper[4759]: E1125 19:33:58.106924 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.106987 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:33:58 crc kubenswrapper[4759]: E1125 19:33:58.107143 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.164550 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.164780 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.164861 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.164922 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.164977 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.267835 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.268086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.268146 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.268224 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.268304 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.371600 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.371642 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.371658 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.371683 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.371700 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.385849 4759 generic.go:334] "Generic (PLEG): container finished" podID="6dff2992-ff1a-4dce-b635-ee0f16ba11d8" containerID="259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22" exitCode=0 Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.385926 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerDied","Data":"259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.390821 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.391437 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.391564 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.399721 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.420370 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.422429 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.428131 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.440087 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.453627 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.464302 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.474438 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.474480 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.474489 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.474502 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.474512 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.480880 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.491853 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.504295 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.516292 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.530009 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.540801 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.550914 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.562417 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576305 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576348 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576363 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576397 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.576929 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.585771 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.594358 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.604841 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.614929 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.623447 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.632509 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.641123 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.651381 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.660534 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.672122 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.683129 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.683938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.683964 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.683982 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.683999 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.684010 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.701779 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.713384 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.736241 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:58Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.786288 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.786331 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.786339 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.786353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.786362 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.888043 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.888409 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.888429 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.888461 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.888506 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.990791 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.990825 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.990835 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.990848 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:58 crc kubenswrapper[4759]: I1125 19:33:58.990858 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:58Z","lastTransitionTime":"2025-11-25T19:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.093923 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.094023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.094046 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.094073 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.094095 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.106575 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:33:59 crc kubenswrapper[4759]: E1125 19:33:59.106733 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.196490 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.196539 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.196561 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.196588 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.196608 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.301611 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.301651 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.301662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.301677 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.301689 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.398264 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" event={"ID":"6dff2992-ff1a-4dce-b635-ee0f16ba11d8","Type":"ContainerStarted","Data":"b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.398457 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.404088 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.404134 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.404148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.404166 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.404181 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.412745 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.429920 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.444163 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.466249 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.478232 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.489259 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.498184 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.505907 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.505970 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.505980 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.506000 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.506009 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.509793 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.520277 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.529849 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.542413 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.553942 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.566805 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.580651 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:33:59Z is after 2025-08-24T17:21:41Z" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.608459 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.608507 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.608518 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.608533 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.608543 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.710576 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.710604 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.710612 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.710623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.710631 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.813687 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.813740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.813758 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.813780 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.813797 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.916694 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.916734 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.916748 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.916765 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:33:59 crc kubenswrapper[4759]: I1125 19:33:59.916777 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:33:59Z","lastTransitionTime":"2025-11-25T19:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.019745 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.019787 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.019802 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.019823 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.019839 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.106817 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:00 crc kubenswrapper[4759]: E1125 19:34:00.106908 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.107164 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:00 crc kubenswrapper[4759]: E1125 19:34:00.107228 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.118335 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.121811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.121856 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.121868 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.121894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.121940 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.128617 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.149400 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.161191 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.191126 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.221528 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.222107 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.227685 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.227719 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.227727 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.227741 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.227754 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.239806 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.253958 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.265921 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.288963 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.301028 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.315168 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.328580 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.330086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.330127 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.330140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.330156 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.330166 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.352255 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.365436 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.382541 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.399513 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.401322 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.422229 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.432002 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.432054 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.432063 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.432079 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.432089 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.434895 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.448329 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.461096 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.473288 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.488302 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.501211 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.516398 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.528002 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.534613 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.534659 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.534667 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.534680 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.534717 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.539129 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.552149 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:00Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.637756 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.637800 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.637811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.637831 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.637841 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.739937 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.739977 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.739985 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.739999 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.740008 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.842394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.842432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.842462 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.842480 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.842491 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.944961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.944996 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.945004 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.945018 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:00 crc kubenswrapper[4759]: I1125 19:34:00.945028 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:00Z","lastTransitionTime":"2025-11-25T19:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.047750 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.047781 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.047789 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.047804 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.047814 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.106094 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:01 crc kubenswrapper[4759]: E1125 19:34:01.106279 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.150330 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.150382 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.150399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.150422 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.150438 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.253409 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.253519 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.253541 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.253570 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.253586 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.356864 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.356908 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.356920 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.356938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.356949 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.407198 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/0.log" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.411785 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc" exitCode=1 Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.411848 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.412918 4759 scope.go:117] "RemoveContainer" containerID="190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.428333 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.456333 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.459052 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.459095 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.459108 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.459126 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.459138 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.471483 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.490752 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.506551 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.534466 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.549399 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.561413 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.561454 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.561476 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.561493 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.561504 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.568038 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.583890 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.608794 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.628973 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.645650 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.664856 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.664897 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.664905 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.664921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.664933 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.665641 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.679688 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:01Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.767666 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.767704 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.767714 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.767729 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.767740 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.870521 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.870590 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.870610 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.870634 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.870652 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.973111 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.973485 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.973761 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.973957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:01 crc kubenswrapper[4759]: I1125 19:34:01.974102 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:01Z","lastTransitionTime":"2025-11-25T19:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.077073 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.077492 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.077511 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.077535 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.077551 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.106675 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.106793 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:02 crc kubenswrapper[4759]: E1125 19:34:02.106914 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:02 crc kubenswrapper[4759]: E1125 19:34:02.107014 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.180834 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.180873 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.180893 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.180911 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.180922 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.282995 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.283030 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.283043 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.283062 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.283073 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.385848 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.385900 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.385913 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.385930 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.385945 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.416611 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/0.log" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.419421 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.419575 4759 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.431384 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.441394 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.451929 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.464838 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.481190 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.487979 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.488556 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.488637 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.488701 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.488778 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.495175 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.509726 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.523628 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.531854 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.543974 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.557572 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.570898 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.581436 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.591860 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.591905 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.591919 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.591936 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.591947 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.600751 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:02Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.694392 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.694660 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.694745 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.694822 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.694892 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.797178 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.797252 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.797275 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.797299 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.797316 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.899698 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.899761 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.899783 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.899812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:02 crc kubenswrapper[4759]: I1125 19:34:02.899834 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:02Z","lastTransitionTime":"2025-11-25T19:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.003558 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.003601 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.003614 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.003630 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.003642 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.105880 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:03 crc kubenswrapper[4759]: E1125 19:34:03.106018 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.106044 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.106085 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.106101 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.106124 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.106140 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.209324 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.209717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.210649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.210910 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.211188 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.314398 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.314432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.314445 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.314481 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.314492 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.418248 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.418310 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.418328 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.418351 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.418367 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.427951 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/1.log" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.429305 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/0.log" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.433132 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1" exitCode=1 Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.433174 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.433227 4759 scope.go:117] "RemoveContainer" containerID="190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.434041 4759 scope.go:117] "RemoveContainer" containerID="555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1" Nov 25 19:34:03 crc kubenswrapper[4759]: E1125 19:34:03.434243 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.460225 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.479404 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.497576 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.513923 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.520832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.520853 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.520862 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.520878 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.520889 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.531420 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.551812 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.562886 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.578853 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.593593 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.608696 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.623681 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.623730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.623742 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.623760 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.623772 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.630329 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.645765 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.660587 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.670829 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.725836 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.725875 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.725884 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.725898 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.725908 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.828360 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.828622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.828717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.828812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.828888 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.931755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.931807 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.931819 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.931839 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.931853 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:03Z","lastTransitionTime":"2025-11-25T19:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.981017 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw"] Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.981784 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.983405 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.983881 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 19:34:03 crc kubenswrapper[4759]: I1125 19:34:03.996925 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:03Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.014759 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.025891 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.033927 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.033975 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.033987 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.034006 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.034019 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.041480 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.056744 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.073063 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.077013 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5ltr\" (UniqueName: \"kubernetes.io/projected/6bd110cd-4afd-4c0a-b5a6-5a1062326603-kube-api-access-z5ltr\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.077137 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.077243 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.077316 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.083591 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.102798 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.105983 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.106059 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:04 crc kubenswrapper[4759]: E1125 19:34:04.106100 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:04 crc kubenswrapper[4759]: E1125 19:34:04.106241 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.120803 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.134746 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.136293 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.136409 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.136430 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.136455 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.136509 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.153284 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.174250 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.178051 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.178139 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5ltr\" (UniqueName: \"kubernetes.io/projected/6bd110cd-4afd-4c0a-b5a6-5a1062326603-kube-api-access-z5ltr\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.178217 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.178264 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.179278 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.179337 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.187345 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd110cd-4afd-4c0a-b5a6-5a1062326603-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.187750 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.200435 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.204310 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5ltr\" (UniqueName: \"kubernetes.io/projected/6bd110cd-4afd-4c0a-b5a6-5a1062326603-kube-api-access-z5ltr\") pod \"ovnkube-control-plane-749d76644c-vczjw\" (UID: \"6bd110cd-4afd-4c0a-b5a6-5a1062326603\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.216673 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:04Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.239578 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.239622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.239636 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.239653 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.239667 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.301686 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" Nov 25 19:34:04 crc kubenswrapper[4759]: W1125 19:34:04.313829 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd110cd_4afd_4c0a_b5a6_5a1062326603.slice/crio-d2b3ec36828527de120bc20c5f8619aa68ff392cfcc2b8dfe23eac724dc3e9bf WatchSource:0}: Error finding container d2b3ec36828527de120bc20c5f8619aa68ff392cfcc2b8dfe23eac724dc3e9bf: Status 404 returned error can't find the container with id d2b3ec36828527de120bc20c5f8619aa68ff392cfcc2b8dfe23eac724dc3e9bf Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.343040 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.343070 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.343081 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.343097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.343110 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.437513 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/1.log" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.442193 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" event={"ID":"6bd110cd-4afd-4c0a-b5a6-5a1062326603","Type":"ContainerStarted","Data":"d2b3ec36828527de120bc20c5f8619aa68ff392cfcc2b8dfe23eac724dc3e9bf"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.444607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.444665 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.444678 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.444697 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.444712 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.547393 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.547432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.547440 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.547498 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.547508 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.652433 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.652542 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.652564 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.652589 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.652606 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.755248 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.755302 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.755315 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.755336 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.755349 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.857364 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.857425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.857437 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.857484 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.857499 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.959764 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.959797 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.959805 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.959819 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:04 crc kubenswrapper[4759]: I1125 19:34:04.959828 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:04Z","lastTransitionTime":"2025-11-25T19:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.062109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.062148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.062156 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.062169 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.062177 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.106218 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.106364 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.165541 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.165899 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.165909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.165924 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.165935 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.267917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.267954 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.267962 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.267975 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.267984 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.370560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.370585 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.370593 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.370607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.370616 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.448395 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" event={"ID":"6bd110cd-4afd-4c0a-b5a6-5a1062326603","Type":"ContainerStarted","Data":"2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.448435 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" event={"ID":"6bd110cd-4afd-4c0a-b5a6-5a1062326603","Type":"ContainerStarted","Data":"a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.449474 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-k5q6v"] Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.449868 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.449923 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.471350 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.472933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.472968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.472978 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.472992 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.473003 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.490900 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.491517 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.491675 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgmgq\" (UniqueName: \"kubernetes.io/projected/8e41227a-a128-4f1e-a21c-2242cea52970-kube-api-access-hgmgq\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.505371 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.518123 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.537643 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.553639 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.569930 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.581184 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.581221 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.581229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.581243 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.581252 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.593874 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.593966 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.594045 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgmgq\" (UniqueName: \"kubernetes.io/projected/8e41227a-a128-4f1e-a21c-2242cea52970-kube-api-access-hgmgq\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.594201 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.594304 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:06.09427476 +0000 UTC m=+36.680222491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.613270 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.619513 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgmgq\" (UniqueName: \"kubernetes.io/projected/8e41227a-a128-4f1e-a21c-2242cea52970-kube-api-access-hgmgq\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.636310 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.652703 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.671608 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.683923 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.683957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.683968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.683983 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.683995 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.689903 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.702018 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.712471 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.721940 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.733406 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.747502 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.757111 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.770831 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.785541 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.786554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.786586 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.786595 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.786613 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.786622 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.800549 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.814990 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.827836 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.839591 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.861099 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.874891 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.889391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.889432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.889443 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.889471 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.889480 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.896295 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.908098 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.924504 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190486c8e0ec22972750d06d28b54d5e4580915f4917484e13283afd5513b2bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"message\\\":\\\".openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00016447b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 19:34:00.626804 6050 services_controller.go:443] Built service openshift-apiserver/check-endpoints LB cluster-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.936205 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:05Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.991157 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.991203 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.991217 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.991238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.991252 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:05Z","lastTransitionTime":"2025-11-25T19:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.997849 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.997958 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.997975 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.997952544 +0000 UTC m=+52.583900255 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.997999 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.998045 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:05 crc kubenswrapper[4759]: I1125 19:34:05.998068 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998146 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998186 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998190 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998203 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998208 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998221 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998241 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998266 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.998243942 +0000 UTC m=+52.584191653 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998287 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.998278983 +0000 UTC m=+52.584226694 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998340 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.998308094 +0000 UTC m=+52.584255855 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998155 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:05 crc kubenswrapper[4759]: E1125 19:34:05.998428 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.998406917 +0000 UTC m=+52.584354728 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.029924 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.029967 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.029976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.029990 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.030000 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.041698 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:06Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.045049 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.045082 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.045093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.045109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.045119 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.058343 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:06Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.062114 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.062173 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.062191 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.062213 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.062230 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.076151 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:06Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.079604 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.079650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.079669 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.079688 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.079703 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.093619 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:06Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.097276 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.097328 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.097338 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.097353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.097364 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.098821 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.098974 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.099036 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:07.099020571 +0000 UTC m=+37.684968272 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.106227 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.106344 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.106572 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.106757 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.110582 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:06Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:06 crc kubenswrapper[4759]: E1125 19:34:06.110691 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.112727 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.112773 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.112789 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.112811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.112828 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.215698 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.215746 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.215765 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.215789 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.215807 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.318755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.318800 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.318814 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.318835 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.318849 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.421928 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.421986 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.421998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.422017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.422030 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.524846 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.524906 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.524917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.524931 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.524941 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.627850 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.627916 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.627933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.627957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.627977 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.730436 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.730528 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.730545 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.730573 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.730597 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.833408 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.833522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.833546 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.833570 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.833587 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.936628 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.936700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.936725 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.936751 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:06 crc kubenswrapper[4759]: I1125 19:34:06.936773 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:06Z","lastTransitionTime":"2025-11-25T19:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.038982 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.039054 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.039077 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.039107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.039131 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.106748 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.106754 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:07 crc kubenswrapper[4759]: E1125 19:34:07.106892 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:07 crc kubenswrapper[4759]: E1125 19:34:07.107009 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.133039 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:07 crc kubenswrapper[4759]: E1125 19:34:07.133182 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:07 crc kubenswrapper[4759]: E1125 19:34:07.133237 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:09.133221328 +0000 UTC m=+39.719169039 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.141690 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.141730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.141743 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.141759 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.141771 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.244283 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.244360 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.244373 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.244391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.244403 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.347285 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.347323 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.347332 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.347346 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.347355 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.449575 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.449614 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.449623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.449638 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.449649 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.552097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.552142 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.552156 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.552173 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.552185 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.654122 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.654209 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.654229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.654251 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.654269 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.756984 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.757037 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.757046 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.757064 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.757074 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.856792 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.857558 4759 scope.go:117] "RemoveContainer" containerID="555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1" Nov 25 19:34:07 crc kubenswrapper[4759]: E1125 19:34:07.857699 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.858931 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.858958 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.858968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.858978 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.858988 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.868825 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.884165 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.896658 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.909570 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.920719 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.941995 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.954751 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.960713 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.960762 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.960775 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.960792 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.960804 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:07Z","lastTransitionTime":"2025-11-25T19:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.966834 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.977161 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:07 crc kubenswrapper[4759]: I1125 19:34:07.992587 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:07Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.006997 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.019494 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.032482 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.046092 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.061447 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.063107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.063133 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.063142 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.063165 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.063176 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.072327 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:08Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.106563 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.106572 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:08 crc kubenswrapper[4759]: E1125 19:34:08.106689 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:08 crc kubenswrapper[4759]: E1125 19:34:08.106757 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.165280 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.165310 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.165320 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.165335 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.165345 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.268031 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.268093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.268105 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.268118 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.268130 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.370378 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.370718 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.370812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.370893 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.370975 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.472788 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.472845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.472867 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.472894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.472917 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.575831 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.575893 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.575910 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.575933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.575949 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.678428 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.678933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.679223 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.679530 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.679811 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.783271 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.783333 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.783353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.783380 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.783400 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.884926 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.884970 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.884982 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.884996 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.885006 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.989836 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.989917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.989942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.989974 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:08 crc kubenswrapper[4759]: I1125 19:34:08.990008 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:08Z","lastTransitionTime":"2025-11-25T19:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.093046 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.093087 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.093099 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.093115 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.093126 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.106504 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.106574 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:09 crc kubenswrapper[4759]: E1125 19:34:09.106810 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:09 crc kubenswrapper[4759]: E1125 19:34:09.106615 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.151744 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:09 crc kubenswrapper[4759]: E1125 19:34:09.151942 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:09 crc kubenswrapper[4759]: E1125 19:34:09.152036 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:13.152014808 +0000 UTC m=+43.737962519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.195717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.195779 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.195797 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.195823 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.195842 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.298799 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.298833 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.298841 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.298855 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.298863 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.402057 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.402148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.402171 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.402692 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.402972 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.506019 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.506079 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.506097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.506120 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.506138 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.608385 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.608478 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.608497 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.608522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.608539 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.711882 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.711931 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.711952 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.711979 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.711998 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.814976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.815021 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.815030 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.815044 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.815056 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.917508 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.917564 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.917584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.917607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:09 crc kubenswrapper[4759]: I1125 19:34:09.917626 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:09Z","lastTransitionTime":"2025-11-25T19:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.021548 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.021604 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.021622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.021644 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.021692 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.105982 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:10 crc kubenswrapper[4759]: E1125 19:34:10.106162 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.106357 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:10 crc kubenswrapper[4759]: E1125 19:34:10.106582 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.126216 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.126310 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.126329 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.126351 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.126368 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.142181 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.175207 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.192881 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.214055 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.228607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.228631 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.228638 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.228650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.228658 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.235357 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.251418 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.269987 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.289089 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.307172 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.323784 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.331882 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.331934 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.331952 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.331977 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.331994 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.344584 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.370428 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.386551 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.408973 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.426395 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.435548 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.435834 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.435981 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.436111 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.436247 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.449836 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:10Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.539111 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.539163 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.539180 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.539204 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.539221 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.642731 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.643077 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.643213 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.643353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.643533 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.747269 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.747319 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.747340 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.747363 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.747381 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.850762 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.850819 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.850837 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.850861 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.850882 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.955222 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.955291 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.955308 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.955332 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:10 crc kubenswrapper[4759]: I1125 19:34:10.955348 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:10Z","lastTransitionTime":"2025-11-25T19:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.057617 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.057669 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.057681 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.057700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.057713 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.106432 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.106500 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:11 crc kubenswrapper[4759]: E1125 19:34:11.106667 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:11 crc kubenswrapper[4759]: E1125 19:34:11.106784 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.161679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.161914 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.161938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.161966 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.161986 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.269437 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.269490 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.269501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.269517 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.269528 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.371986 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.372034 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.372046 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.372064 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.372076 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.474388 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.474493 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.474514 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.474538 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.474556 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.576984 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.577037 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.577054 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.577075 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.577092 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.679866 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.679915 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.679927 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.679944 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.679956 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.782228 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.782268 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.782277 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.782290 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.782300 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.885016 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.885070 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.885088 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.885110 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.885126 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.988066 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.988396 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.988525 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.988653 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:11 crc kubenswrapper[4759]: I1125 19:34:11.988763 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:11Z","lastTransitionTime":"2025-11-25T19:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.092071 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.092115 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.092124 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.092139 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.092148 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.106146 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.106172 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:12 crc kubenswrapper[4759]: E1125 19:34:12.106235 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:12 crc kubenswrapper[4759]: E1125 19:34:12.106345 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.194816 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.195332 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.195428 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.195524 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.195607 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.297924 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.297987 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.298009 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.298049 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.298072 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.400899 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.400948 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.400956 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.400970 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.400978 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.503095 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.503143 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.503151 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.503164 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.503173 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.606083 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.606123 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.606131 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.606145 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.606154 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.708635 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.708673 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.708683 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.708698 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.708710 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.811410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.811483 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.811501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.811524 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.811542 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.913631 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.913688 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.913707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.913745 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:12 crc kubenswrapper[4759]: I1125 19:34:12.913783 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:12Z","lastTransitionTime":"2025-11-25T19:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.015885 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.015995 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.016017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.016041 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.016057 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.106562 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.106662 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:13 crc kubenswrapper[4759]: E1125 19:34:13.106680 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:13 crc kubenswrapper[4759]: E1125 19:34:13.106867 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.119308 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.119395 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.119422 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.119494 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.119520 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.189000 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:13 crc kubenswrapper[4759]: E1125 19:34:13.189205 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:13 crc kubenswrapper[4759]: E1125 19:34:13.189274 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:21.189251541 +0000 UTC m=+51.775199282 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.222266 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.223163 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.223362 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.223602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.223813 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.327052 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.327424 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.327684 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.328158 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.328608 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.432501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.432862 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.433057 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.433248 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.433423 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.536486 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.536545 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.536568 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.536594 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.536614 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.638752 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.638783 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.638794 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.638809 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.638819 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.742535 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.742601 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.742623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.742652 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.742674 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.845861 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.845930 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.845952 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.845981 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.846002 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.948771 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.948884 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.948905 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.948976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:13 crc kubenswrapper[4759]: I1125 19:34:13.949003 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:13Z","lastTransitionTime":"2025-11-25T19:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.051554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.051610 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.051622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.051637 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.051649 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.106532 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:14 crc kubenswrapper[4759]: E1125 19:34:14.106738 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.106811 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:14 crc kubenswrapper[4759]: E1125 19:34:14.106947 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.154622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.154983 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.155135 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.155270 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.155399 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.258110 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.258150 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.258161 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.258176 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.258187 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.360865 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.360921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.360939 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.360961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.360982 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.463806 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.463868 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.463889 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.463915 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.463935 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.566602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.566717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.566737 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.566761 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.566779 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.669116 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.669162 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.669174 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.669196 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.669212 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.772679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.772764 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.772788 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.772815 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.772846 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.876439 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.876560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.876583 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.876611 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.876636 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.979132 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.979186 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.979203 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.979227 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:14 crc kubenswrapper[4759]: I1125 19:34:14.979243 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:14Z","lastTransitionTime":"2025-11-25T19:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.085868 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.085951 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.085979 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.086009 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.086043 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.106582 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.106608 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:15 crc kubenswrapper[4759]: E1125 19:34:15.106740 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:15 crc kubenswrapper[4759]: E1125 19:34:15.106920 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.189984 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.190027 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.190039 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.190058 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.190071 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.292623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.292690 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.292708 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.292733 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.292750 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.396505 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.396562 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.396578 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.396638 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.396692 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.498969 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.499018 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.499035 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.499059 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.499076 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.602225 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.602292 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.602324 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.602351 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.602368 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.705436 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.705536 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.705553 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.705578 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.705595 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.808432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.808501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.808514 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.808533 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.808546 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.911117 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.911169 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.911187 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.911209 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:15 crc kubenswrapper[4759]: I1125 19:34:15.911226 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:15Z","lastTransitionTime":"2025-11-25T19:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.014388 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.014481 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.014502 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.014531 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.014590 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.106515 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.106541 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.106939 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.107139 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.116875 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.116928 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.116950 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.116974 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.116992 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.188981 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.189058 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.189078 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.189104 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.189121 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.209240 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:16Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.214775 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.214844 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.214867 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.214894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.214915 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.235674 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:16Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.240712 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.240766 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.240786 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.240809 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.240827 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.261062 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:16Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.266092 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.266141 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.266153 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.266174 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.266190 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.284557 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:16Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.288977 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.289010 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.289022 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.289036 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.289046 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.309783 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:16Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:16 crc kubenswrapper[4759]: E1125 19:34:16.310007 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.312040 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.312087 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.312104 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.312126 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.312145 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.414880 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.415216 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.415349 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.415512 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.415662 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.518997 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.519036 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.519048 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.519062 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.519073 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.622527 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.622575 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.622585 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.622602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.622615 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.725695 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.725742 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.725753 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.725772 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.725785 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.829502 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.829565 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.829583 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.829609 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.829629 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.933538 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.933640 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.933661 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.933686 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:16 crc kubenswrapper[4759]: I1125 19:34:16.933704 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:16Z","lastTransitionTime":"2025-11-25T19:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.036483 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.036520 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.036529 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.036544 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.036553 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.106520 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.106520 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:17 crc kubenswrapper[4759]: E1125 19:34:17.106837 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:17 crc kubenswrapper[4759]: E1125 19:34:17.106698 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.139888 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.139947 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.139964 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.139985 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.140003 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.243567 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.243707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.243739 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.243768 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.243790 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.347217 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.347287 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.347308 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.347338 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.347362 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.451083 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.451142 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.451159 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.451183 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.451199 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.554827 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.554874 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.554886 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.554901 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.554911 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.657438 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.657527 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.657545 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.657571 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.657589 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.761735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.761842 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.761861 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.761884 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.761901 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.865517 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.865596 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.865614 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.865648 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.865678 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.969121 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.969166 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.969178 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.969196 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:17 crc kubenswrapper[4759]: I1125 19:34:17.969209 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:17Z","lastTransitionTime":"2025-11-25T19:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.072566 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.072958 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.073090 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.073229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.073362 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.105878 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:18 crc kubenswrapper[4759]: E1125 19:34:18.105978 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.106224 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:18 crc kubenswrapper[4759]: E1125 19:34:18.106579 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.177106 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.177143 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.177154 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.177171 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.177182 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.279420 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.279959 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.280068 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.280153 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.280230 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.382851 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.382892 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.382901 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.382915 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.382927 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.486245 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.486324 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.486340 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.486366 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.486385 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.588630 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.588663 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.588671 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.588687 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.588696 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.690313 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.690654 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.690737 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.690807 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.690865 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.812894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.812932 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.812943 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.812961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.812973 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.915612 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.915993 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.916138 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.916280 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:18 crc kubenswrapper[4759]: I1125 19:34:18.916520 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:18Z","lastTransitionTime":"2025-11-25T19:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.019991 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.020389 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.020609 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.020757 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.020898 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.105875 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:19 crc kubenswrapper[4759]: E1125 19:34:19.106043 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.105875 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:19 crc kubenswrapper[4759]: E1125 19:34:19.106224 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.128125 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.128181 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.128197 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.128219 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.128234 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.231742 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.231794 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.231806 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.231847 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.231865 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.334607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.334684 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.334702 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.334725 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.334743 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.437769 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.437829 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.437846 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.437871 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.437888 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.540086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.540139 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.540158 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.540182 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.540198 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.643122 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.643194 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.643227 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.643260 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.643282 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.746889 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.746950 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.746972 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.746996 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.747014 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.850495 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.850560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.850579 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.850607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.850627 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.954296 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.954709 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.954854 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.955037 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:19 crc kubenswrapper[4759]: I1125 19:34:19.955187 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:19Z","lastTransitionTime":"2025-11-25T19:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.058286 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.058735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.058926 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.059111 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.059272 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.106665 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:20 crc kubenswrapper[4759]: E1125 19:34:20.106909 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.106994 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:20 crc kubenswrapper[4759]: E1125 19:34:20.107197 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.126176 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.140031 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.155241 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.162308 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.162629 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.162745 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.163070 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.163197 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.174203 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.189060 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.213778 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.230796 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.246203 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.260876 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.265361 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.265415 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.265427 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.265468 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.265482 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.273386 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.289938 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.301430 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.314210 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.327565 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.349697 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.361216 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.367791 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.367847 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.367864 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.367884 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.367900 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.446603 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.456734 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.459037 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.469950 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.469968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.469976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.469988 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.469996 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.471084 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.481007 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.490675 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.503716 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.516421 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.535156 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.544615 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.553987 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.566116 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.571937 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.571966 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.571974 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.571987 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.571996 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.577738 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.590853 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.601340 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.615569 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.625981 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.636867 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:20Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.674498 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.674525 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.674535 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.674552 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.674563 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.776470 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.776504 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.776513 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.776525 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.776534 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.878438 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.878491 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.878501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.878514 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.878526 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.980245 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.980276 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.980284 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.980296 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:20 crc kubenswrapper[4759]: I1125 19:34:20.980305 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:20Z","lastTransitionTime":"2025-11-25T19:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.082998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.083041 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.083052 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.083067 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.083079 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.106381 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.106405 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:21 crc kubenswrapper[4759]: E1125 19:34:21.106500 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:21 crc kubenswrapper[4759]: E1125 19:34:21.106809 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.107269 4759 scope.go:117] "RemoveContainer" containerID="555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.189407 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.189733 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.189743 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.189756 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.189766 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: E1125 19:34:21.281609 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:21 crc kubenswrapper[4759]: E1125 19:34:21.281726 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:37.281699722 +0000 UTC m=+67.867647463 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.281398 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.292814 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.292873 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.292895 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.292925 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.292946 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.395100 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.395140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.395149 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.395165 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.395176 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.497925 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.497965 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.497976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.497993 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.498005 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.502365 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/1.log" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.506818 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.507262 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.522016 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.541397 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.561087 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.570252 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.594679 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.600509 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.600544 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.600554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.600572 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.600582 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.608960 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.624104 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.643791 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.654876 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.670615 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.685149 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.696099 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.702811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.702854 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.702866 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.702883 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.702895 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.707584 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.720015 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.739960 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.750057 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.764467 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:21Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.805368 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.805409 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.805420 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.805436 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.805466 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.907921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.907967 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.907981 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.908000 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:21 crc kubenswrapper[4759]: I1125 19:34:21.908014 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:21Z","lastTransitionTime":"2025-11-25T19:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.010896 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.010936 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.010947 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.010961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.010985 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.090543 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.090746 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:34:54.090721006 +0000 UTC m=+84.676668697 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.091129 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.091178 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.091301 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091246 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.091339 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091353 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:54.091347103 +0000 UTC m=+84.677294804 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091349 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091385 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091399 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091462 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091474 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:54.091442825 +0000 UTC m=+84.677390566 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091484 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091497 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091518 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091536 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:54.091522078 +0000 UTC m=+84.677469799 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.091596 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:34:54.091575409 +0000 UTC m=+84.677523140 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.106079 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.106112 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.106185 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.106279 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.112909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.112959 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.112975 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.112997 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.113015 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.215839 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.215890 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.215901 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.215917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.215928 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.318333 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.318380 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.318391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.318410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.318424 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.420627 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.420680 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.420696 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.420716 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.420728 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.511183 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/2.log" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.511827 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/1.log" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.513952 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" exitCode=1 Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.513989 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.514034 4759 scope.go:117] "RemoveContainer" containerID="555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.519581 4759 scope.go:117] "RemoveContainer" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" Nov 25 19:34:22 crc kubenswrapper[4759]: E1125 19:34:22.519912 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.525159 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.525188 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.525196 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.525209 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.525220 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.535264 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.547157 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.558589 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.575902 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555ed32434d6b6fc1f726799d3bbf1f09652b519d473863918584d837b7cfae1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1125 19:34:02.556243 6193 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 19:34:02.556258 6193 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 19:34:02.556264 6193 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:02.556293 6193 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:02.557727 6193 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 19:34:02.557743 6193 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 19:34:02.557751 6193 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:02.557759 6193 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 19:34:02.557767 6193 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 19:34:02.557929 6193 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 19:34:02.558027 6193 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 19:34:02.558055 6193 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 19:34:02.558108 6193 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 19:34:02.558216 6193 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 19:34:02.558276 6193 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 19:34:02.558335 6193 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.587088 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.598776 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.622035 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.627340 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.627377 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.627389 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.627406 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.627417 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.643742 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.653274 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.668166 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.679079 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.690794 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.702404 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.711385 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.721769 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.730203 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.730239 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.730249 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.730265 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.730275 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.731190 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.743303 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:22Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.832383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.832421 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.832429 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.832441 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.832466 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.934811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.934844 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.934852 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.934864 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:22 crc kubenswrapper[4759]: I1125 19:34:22.934874 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:22Z","lastTransitionTime":"2025-11-25T19:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.037476 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.037532 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.037545 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.037564 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.037577 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.106609 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.106675 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:23 crc kubenswrapper[4759]: E1125 19:34:23.106767 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:23 crc kubenswrapper[4759]: E1125 19:34:23.106848 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.140353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.140410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.140481 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.140502 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.140517 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.268313 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.268486 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.268504 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.268528 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.268545 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.371016 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.371064 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.371076 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.371093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.371103 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.473499 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.473551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.473567 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.473588 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.473604 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.517731 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/2.log" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.520776 4759 scope.go:117] "RemoveContainer" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" Nov 25 19:34:23 crc kubenswrapper[4759]: E1125 19:34:23.520924 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.534787 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.551783 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.567165 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.575854 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.575910 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.575925 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.575946 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.575960 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.581339 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.596703 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.610391 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.624657 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.638539 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.649016 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.663960 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.678909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.678956 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.678972 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.678995 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.679015 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.676839 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.696042 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.710367 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.732593 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.743164 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.752801 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.762720 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:23Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.781219 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.781258 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.781270 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.781287 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.781299 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.883097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.883138 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.883148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.883162 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.883171 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.985894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.986179 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.986287 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.986425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:23 crc kubenswrapper[4759]: I1125 19:34:23.986542 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:23Z","lastTransitionTime":"2025-11-25T19:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.089163 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.089199 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.089208 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.089222 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.089232 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.106818 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.107033 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:24 crc kubenswrapper[4759]: E1125 19:34:24.107138 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:24 crc kubenswrapper[4759]: E1125 19:34:24.107318 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.191584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.191619 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.191628 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.191643 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.191653 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.294868 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.295117 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.295181 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.295267 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.295349 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.398646 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.398702 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.398724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.398750 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.398763 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.501602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.501650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.501662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.501679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.501690 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.604059 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.604109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.604123 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.604142 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.604154 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.706755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.706799 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.706816 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.706835 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.706851 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.809662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.809717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.809740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.809768 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.809789 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.912395 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.912435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.912482 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.912498 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:24 crc kubenswrapper[4759]: I1125 19:34:24.912511 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:24Z","lastTransitionTime":"2025-11-25T19:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.015112 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.015145 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.015172 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.015186 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.015195 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.106865 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.106877 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:25 crc kubenswrapper[4759]: E1125 19:34:25.107063 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:25 crc kubenswrapper[4759]: E1125 19:34:25.107197 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.118200 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.118282 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.118334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.118357 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.118377 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.221659 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.221707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.221723 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.221749 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.221767 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.324257 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.324299 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.324307 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.324321 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.324330 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.426776 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.426817 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.426830 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.426848 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.426864 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.528492 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.528537 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.528550 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.528565 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.528575 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.630305 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.630353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.630364 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.630383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.630394 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.736780 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.736830 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.736842 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.736859 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.736871 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.839471 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.839513 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.839532 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.839554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.839569 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.942578 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.942628 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.942650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.942678 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:25 crc kubenswrapper[4759]: I1125 19:34:25.942701 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:25Z","lastTransitionTime":"2025-11-25T19:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.045348 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.045413 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.045435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.045495 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.045519 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.106155 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.106180 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.106292 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.106386 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.147287 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.147338 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.147354 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.147376 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.147391 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.250060 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.250107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.250123 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.250141 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.250153 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.353107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.353158 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.353171 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.353187 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.353197 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.456350 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.456405 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.456420 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.456478 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.456499 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.559085 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.559112 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.559120 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.559132 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.559142 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.661730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.661799 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.661811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.661828 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.661840 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.689352 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.689383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.689394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.689408 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.689418 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.700210 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:26Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.703265 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.703315 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.703334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.703356 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.703373 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.718352 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:26Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.722368 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.722396 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.722409 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.722425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.722436 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.737915 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:26Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.740847 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.740894 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.740906 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.740922 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.740937 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.756154 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:26Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.759832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.759862 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.759870 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.759885 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.759894 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.770200 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:26Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:26 crc kubenswrapper[4759]: E1125 19:34:26.770308 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.771804 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.771827 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.771835 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.771846 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.771855 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.873345 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.873369 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.873377 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.873389 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.873397 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.975340 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.975369 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.975379 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.975390 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:26 crc kubenswrapper[4759]: I1125 19:34:26.975398 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:26Z","lastTransitionTime":"2025-11-25T19:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.077856 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.077907 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.077918 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.077936 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.077946 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.106389 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.106438 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:27 crc kubenswrapper[4759]: E1125 19:34:27.106527 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:27 crc kubenswrapper[4759]: E1125 19:34:27.106583 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.181285 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.181490 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.181522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.181546 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.181562 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.284693 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.284783 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.284797 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.284816 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.284828 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.387685 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.387726 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.387735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.387749 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.387758 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.491760 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.491832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.491856 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.491890 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.491921 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.597416 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.597511 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.597530 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.597561 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.597583 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.700506 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.700561 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.700580 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.700607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.700629 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.803980 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.804044 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.804060 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.804089 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.804106 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.906542 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.906571 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.906579 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.906592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:27 crc kubenswrapper[4759]: I1125 19:34:27.906618 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:27Z","lastTransitionTime":"2025-11-25T19:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.009067 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.009104 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.009114 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.009129 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.009138 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.106488 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.106623 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:28 crc kubenswrapper[4759]: E1125 19:34:28.106791 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:28 crc kubenswrapper[4759]: E1125 19:34:28.106945 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.111394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.111426 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.111435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.111463 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.111472 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.214015 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.214081 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.214099 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.214121 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.214139 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.317898 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.317957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.317975 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.317998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.318014 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.421041 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.421138 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.421157 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.421181 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.421199 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.523858 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.523925 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.523942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.523968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.523986 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.627074 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.627123 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.627133 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.627151 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.627163 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.730400 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.730481 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.730499 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.730522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.730540 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.834238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.834307 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.834331 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.834358 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.834378 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.937322 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.937383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.937405 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.937433 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:28 crc kubenswrapper[4759]: I1125 19:34:28.937496 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:28Z","lastTransitionTime":"2025-11-25T19:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.040973 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.041021 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.041037 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.041059 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.041076 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.106435 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.106442 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:29 crc kubenswrapper[4759]: E1125 19:34:29.106675 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:29 crc kubenswrapper[4759]: E1125 19:34:29.106769 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.143641 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.143696 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.143713 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.143735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.143752 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.245808 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.245854 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.245871 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.245896 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.245912 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.348581 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.348632 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.348649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.348674 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.348691 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.451864 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.451902 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.451913 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.451929 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.451940 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.554499 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.554534 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.554546 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.554567 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.554581 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.657083 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.657151 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.657174 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.657202 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.657224 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.760086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.760135 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.760150 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.760167 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.760181 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.863820 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.863904 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.863929 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.863960 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.863984 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.967644 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.967707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.967725 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.967755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:29 crc kubenswrapper[4759]: I1125 19:34:29.967778 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:29Z","lastTransitionTime":"2025-11-25T19:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.069399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.069439 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.069477 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.069497 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.069511 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.106325 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:30 crc kubenswrapper[4759]: E1125 19:34:30.106568 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.106910 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:30 crc kubenswrapper[4759]: E1125 19:34:30.107061 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.122030 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.135571 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.156908 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.172383 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.172465 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.172478 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.172493 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.172506 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.178299 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.199294 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.229594 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.243404 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.256536 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.266740 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.274933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.274959 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.274972 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.274989 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.275002 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.278780 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.294285 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.311170 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.326434 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.343206 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.358984 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.376481 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.378174 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.378224 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.378264 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.378284 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.378297 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.387581 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:30Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.480353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.480394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.480406 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.480425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.480438 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.583379 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.583412 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.583422 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.583436 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.583458 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.686002 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.686055 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.686077 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.686097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.686142 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.789044 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.789097 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.789113 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.789135 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.789152 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.892034 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.892075 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.892086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.892102 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.892111 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.994818 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.994882 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.994893 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.994928 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:30 crc kubenswrapper[4759]: I1125 19:34:30.994943 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:30Z","lastTransitionTime":"2025-11-25T19:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.097289 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.097354 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.097370 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.097394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.097414 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.106632 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.106713 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:31 crc kubenswrapper[4759]: E1125 19:34:31.106783 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:31 crc kubenswrapper[4759]: E1125 19:34:31.106935 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.200627 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.200679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.200696 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.200720 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.200738 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.304540 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.304647 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.304709 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.304737 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.304758 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.407022 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.407096 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.407118 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.407140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.407158 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.511565 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.511643 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.511662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.511688 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.511723 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.614746 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.614790 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.614799 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.614814 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.614824 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.717824 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.717889 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.717913 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.717941 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.717959 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.820870 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.820938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.820957 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.820982 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.821001 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.924243 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.924297 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.924313 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.924336 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:31 crc kubenswrapper[4759]: I1125 19:34:31.924352 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:31Z","lastTransitionTime":"2025-11-25T19:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.028937 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.029009 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.029028 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.029056 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.029079 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.106087 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.106170 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:32 crc kubenswrapper[4759]: E1125 19:34:32.106360 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:32 crc kubenswrapper[4759]: E1125 19:34:32.106632 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.131579 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.131655 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.131682 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.131721 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.131747 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.234732 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.234823 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.234852 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.234883 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.234905 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.338086 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.338121 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.338129 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.338146 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.338155 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.440717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.440777 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.440800 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.440829 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.440846 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.544337 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.544399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.544422 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.544485 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.544517 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.647596 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.647656 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.647673 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.647700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.647717 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.750551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.750601 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.750619 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.750642 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.750658 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.854876 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.854928 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.854944 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.854966 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.854979 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.958986 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.959033 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.959046 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.959066 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:32 crc kubenswrapper[4759]: I1125 19:34:32.959083 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:32Z","lastTransitionTime":"2025-11-25T19:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.061724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.061794 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.061814 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.061832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.061845 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.106841 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.106937 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:33 crc kubenswrapper[4759]: E1125 19:34:33.106995 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:33 crc kubenswrapper[4759]: E1125 19:34:33.107117 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.165008 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.165073 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.165092 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.165118 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.165136 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.268597 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.268636 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.268647 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.268662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.268671 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.371807 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.371845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.371853 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.371865 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.371875 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.475648 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.475727 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.475753 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.475798 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.475821 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.579592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.579701 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.579743 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.579776 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.579802 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.682693 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.682729 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.682738 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.682753 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.682763 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.785458 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.785501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.785530 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.785547 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.785559 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.889354 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.889395 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.889405 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.889421 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.889430 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.992587 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.992630 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.992644 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.992658 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:33 crc kubenswrapper[4759]: I1125 19:34:33.992669 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:33Z","lastTransitionTime":"2025-11-25T19:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.095551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.095640 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.095668 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.095707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.095734 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.106139 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.106215 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:34 crc kubenswrapper[4759]: E1125 19:34:34.106326 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:34 crc kubenswrapper[4759]: E1125 19:34:34.106478 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.197829 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.197868 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.197878 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.197895 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.197905 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.301148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.301191 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.301203 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.301219 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.301228 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.403845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.403889 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.403897 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.403912 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.403923 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.506627 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.506682 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.506698 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.506722 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.506739 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.609029 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.609063 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.609072 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.609087 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.609098 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.711727 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.711760 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.711770 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.711785 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.711794 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.813818 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.813845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.813853 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.813866 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.813875 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.916953 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.917011 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.917022 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.917040 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:34 crc kubenswrapper[4759]: I1125 19:34:34.917086 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:34Z","lastTransitionTime":"2025-11-25T19:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.021327 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.021366 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.021374 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.021391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.021400 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.106595 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.106680 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:35 crc kubenswrapper[4759]: E1125 19:34:35.106721 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:35 crc kubenswrapper[4759]: E1125 19:34:35.106805 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.123471 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.123522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.123536 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.123554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.123567 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.226679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.226716 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.226726 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.226768 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.226778 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.329153 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.329202 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.329212 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.329226 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.329236 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.432337 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.432372 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.432380 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.432394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.432402 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.535102 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.535141 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.535151 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.535168 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.535179 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.638008 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.638053 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.638066 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.638082 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.638092 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.740908 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.740944 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.740952 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.740965 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.740976 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.842295 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.842333 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.842342 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.842356 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.842364 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.944246 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.944290 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.944301 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.944320 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:35 crc kubenswrapper[4759]: I1125 19:34:35.944333 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:35Z","lastTransitionTime":"2025-11-25T19:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.046387 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.046437 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.046470 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.046489 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.046503 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.106321 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.106338 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.106430 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.106522 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.107057 4759 scope.go:117] "RemoveContainer" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.107273 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.148886 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.148919 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.148927 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.148938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.148947 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.250617 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.250669 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.250688 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.250713 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.250730 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.353963 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.354016 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.354033 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.354056 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.354074 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.456942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.456998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.457017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.457044 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.457069 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.560560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.560603 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.560616 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.560634 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.560646 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.663669 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.663739 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.663756 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.663779 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.663794 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.766222 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.766264 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.766276 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.766293 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.766305 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.868979 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.869034 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.869053 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.869075 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.869092 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.944705 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.944740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.944748 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.944760 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.944768 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.959864 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:36Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.962416 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.962461 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.962470 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.962484 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.962492 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.975465 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:36Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.978961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.979008 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.979020 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.979034 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.979044 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:36 crc kubenswrapper[4759]: E1125 19:34:36.993079 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:36Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.996140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.996176 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.996185 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.996200 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:36 crc kubenswrapper[4759]: I1125 19:34:36.996210 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:36Z","lastTransitionTime":"2025-11-25T19:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.007614 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:37Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.010361 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.010387 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.010395 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.010407 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.010417 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.022791 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:37Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.022899 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.024201 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.024225 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.024233 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.024246 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.024255 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.106168 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.106263 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.106171 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.106565 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.126535 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.126569 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.126581 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.126596 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.126609 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.228464 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.228496 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.228505 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.228517 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.228526 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.330998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.331054 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.331073 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.331099 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.331117 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.355464 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.355602 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:37 crc kubenswrapper[4759]: E1125 19:34:37.355751 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:35:09.355670188 +0000 UTC m=+99.941617899 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.433488 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.433549 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.433562 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.433581 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.433594 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.536251 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.536435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.536519 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.536554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.536572 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.638967 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.639007 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.639017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.639032 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.639041 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.741784 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.741878 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.741897 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.741921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.741940 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.844263 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.844331 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.844356 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.844396 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.844413 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.946255 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.946283 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.946290 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.946303 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:37 crc kubenswrapper[4759]: I1125 19:34:37.946311 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:37Z","lastTransitionTime":"2025-11-25T19:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.048205 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.048238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.048246 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.048262 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.048271 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.106616 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.106647 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:38 crc kubenswrapper[4759]: E1125 19:34:38.106875 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:38 crc kubenswrapper[4759]: E1125 19:34:38.107024 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.150431 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.150486 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.150496 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.150510 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.150520 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.253426 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.253574 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.253584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.253620 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.253629 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.356513 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.356560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.356571 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.356588 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.356599 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.459391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.459428 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.459461 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.459477 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.459486 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.561876 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.561909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.561918 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.561933 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.561941 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.568072 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/0.log" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.568114 4759 generic.go:334] "Generic (PLEG): container finished" podID="67e7b861-9f72-41ba-ab98-35e0a2bdbe39" containerID="105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6" exitCode=1 Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.568140 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerDied","Data":"105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.568508 4759 scope.go:117] "RemoveContainer" containerID="105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.583833 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.598003 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.613621 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.629664 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.642241 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.653893 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.663252 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.664023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.664258 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.664335 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.664402 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.664484 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.683166 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.694306 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.709800 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.723006 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.734783 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.750964 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.761409 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.767028 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.767062 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.767074 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.767107 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.767124 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.773638 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.782652 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.790316 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:38Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.869353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.869417 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.869430 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.869487 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.869503 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.972634 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.972679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.972689 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.972704 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:38 crc kubenswrapper[4759]: I1125 19:34:38.972715 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:38Z","lastTransitionTime":"2025-11-25T19:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.075424 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.075496 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.075508 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.075525 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.075539 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.106296 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.106349 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:39 crc kubenswrapper[4759]: E1125 19:34:39.106441 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:39 crc kubenswrapper[4759]: E1125 19:34:39.106538 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.177394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.177705 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.177782 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.177862 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.177918 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.280512 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.280546 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.280555 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.280570 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.280582 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.382775 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.382814 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.382824 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.382837 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.382847 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.484571 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.484607 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.484615 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.484633 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.484643 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.579490 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/0.log" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.579557 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerStarted","Data":"224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.587024 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.587126 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.587184 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.587250 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.587313 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.600919 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.615461 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.631506 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.661900 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.674672 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.684487 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.689799 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.689836 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.689845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.689860 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.689870 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.696614 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.706513 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.719415 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.728889 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.741123 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.751063 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.761021 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.774656 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.786500 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.791473 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.791522 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.791538 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.791558 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.791573 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.797615 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.811310 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:39Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.893851 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.893897 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.893906 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.893921 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.893934 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.996735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.996779 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.996788 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.996802 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:39 crc kubenswrapper[4759]: I1125 19:34:39.996812 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:39Z","lastTransitionTime":"2025-11-25T19:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.099932 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.100281 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.100437 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.100752 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.100903 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.106608 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:40 crc kubenswrapper[4759]: E1125 19:34:40.106704 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.106785 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:40 crc kubenswrapper[4759]: E1125 19:34:40.106931 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.118667 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.120285 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.131078 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.144349 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.162349 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.180982 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.194785 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.202218 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.202250 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.202259 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.202272 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.202281 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.208638 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.223473 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.236903 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.250774 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.264570 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.282237 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.302362 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.304439 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.304491 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.304500 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.304515 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.304525 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.316895 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.329952 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.351029 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.363721 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:40Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.408164 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.408207 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.408218 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.408234 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.408246 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.510811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.510931 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.510953 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.510985 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.511010 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.614692 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.614738 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.614752 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.614769 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.614781 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.716953 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.716991 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.717002 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.717018 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.717027 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.819296 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.819348 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.819361 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.819384 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.819398 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.921519 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.921552 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.921560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.921574 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:40 crc kubenswrapper[4759]: I1125 19:34:40.921583 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:40Z","lastTransitionTime":"2025-11-25T19:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.024351 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.024424 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.024479 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.024516 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.024537 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.106142 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:41 crc kubenswrapper[4759]: E1125 19:34:41.106277 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.106401 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:41 crc kubenswrapper[4759]: E1125 19:34:41.106688 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.127582 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.127636 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.127647 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.127662 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.127674 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.230659 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.230765 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.230781 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.230811 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.230831 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.333508 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.333582 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.333592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.333608 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.333623 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.437432 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.437492 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.437501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.437515 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.437526 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.540619 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.540652 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.540663 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.540675 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.540683 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.642562 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.642587 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.642596 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.642608 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.642616 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.745278 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.745335 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.745351 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.745372 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.745387 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.847673 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.847724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.847735 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.847754 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.847766 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.950794 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.950840 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.950848 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.950865 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:41 crc kubenswrapper[4759]: I1125 19:34:41.950876 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:41Z","lastTransitionTime":"2025-11-25T19:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.052899 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.053170 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.053179 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.053191 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.053200 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.105940 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:42 crc kubenswrapper[4759]: E1125 19:34:42.106036 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.106188 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:42 crc kubenswrapper[4759]: E1125 19:34:42.106232 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.154745 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.154793 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.154803 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.154821 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.154835 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.257926 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.257999 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.258032 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.258062 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.258083 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.359961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.359988 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.359997 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.360009 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.360017 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.462334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.462369 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.462377 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.462391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.462400 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.564021 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.564085 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.564105 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.564131 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.564151 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.666722 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.666790 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.666812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.666836 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.666854 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.768399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.768664 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.768689 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.768713 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.768731 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.871026 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.871068 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.871078 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.871126 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.871138 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.973542 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.973588 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.973600 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.973617 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:42 crc kubenswrapper[4759]: I1125 19:34:42.973630 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:42Z","lastTransitionTime":"2025-11-25T19:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.076362 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.076416 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.076435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.076492 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.076516 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.106749 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.106772 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:43 crc kubenswrapper[4759]: E1125 19:34:43.106838 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:43 crc kubenswrapper[4759]: E1125 19:34:43.106940 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.179996 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.180040 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.180049 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.180061 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.180070 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.282691 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.282818 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.282836 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.282861 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.282879 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.385543 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.385574 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.385584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.385597 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.385605 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.488105 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.488188 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.488213 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.488245 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.488270 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.590912 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.590940 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.590948 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.590961 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.590989 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.694601 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.694672 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.694694 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.694725 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.694747 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.797342 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.797641 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.797772 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.797926 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.798034 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.900355 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.900417 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.900430 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.900468 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:43 crc kubenswrapper[4759]: I1125 19:34:43.900482 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:43Z","lastTransitionTime":"2025-11-25T19:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.003344 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.003393 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.003412 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.003431 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.003468 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.105784 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:44 crc kubenswrapper[4759]: E1125 19:34:44.106308 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.105891 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:44 crc kubenswrapper[4759]: E1125 19:34:44.106553 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.105820 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.106709 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.106775 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.106849 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.106938 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.209394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.209434 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.209468 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.209485 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.209498 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.312621 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.312658 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.312667 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.312680 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.312689 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.415555 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.415623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.415647 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.415680 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.415701 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.518229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.518287 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.518306 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.518330 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.518347 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.621642 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.621707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.621730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.621761 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.621784 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.724302 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.724359 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.724376 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.724397 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.724414 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.827093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.827224 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.827248 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.827278 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.827301 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.931035 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.931106 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.931128 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.931157 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:44 crc kubenswrapper[4759]: I1125 19:34:44.931178 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:44Z","lastTransitionTime":"2025-11-25T19:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.034625 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.034697 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.034715 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.034740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.034759 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.106752 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.106856 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:45 crc kubenswrapper[4759]: E1125 19:34:45.106971 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:45 crc kubenswrapper[4759]: E1125 19:34:45.107080 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.140694 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.140755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.140769 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.140788 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.140805 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.244563 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.244628 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.244647 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.244675 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.244692 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.348070 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.348147 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.348167 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.348192 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.348210 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.450857 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.450940 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.450962 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.450994 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.451061 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.554088 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.554157 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.554183 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.554214 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.554243 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.657268 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.657323 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.657339 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.657361 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.657378 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.760532 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.760603 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.760627 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.760655 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.760677 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.864023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.864091 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.864113 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.864141 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.864161 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.967302 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.967374 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.967401 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.967425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:45 crc kubenswrapper[4759]: I1125 19:34:45.967476 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:45Z","lastTransitionTime":"2025-11-25T19:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.069542 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.069608 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.069626 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.069651 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.069668 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.106590 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:46 crc kubenswrapper[4759]: E1125 19:34:46.106749 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.107102 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:46 crc kubenswrapper[4759]: E1125 19:34:46.107197 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.172622 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.172693 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.172717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.172746 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.172771 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.276180 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.276243 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.276263 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.276286 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.276304 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.378881 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.378944 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.378968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.379000 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.379040 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.483154 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.483223 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.483239 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.483264 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.483280 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.586434 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.586551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.586566 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.586584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.586598 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.689238 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.689309 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.689326 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.689350 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.689370 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.792619 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.792879 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.792893 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.792917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.792932 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.895195 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.895252 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.895267 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.895284 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.895294 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.997495 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.997551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.997566 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.997592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:46 crc kubenswrapper[4759]: I1125 19:34:46.997606 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:46Z","lastTransitionTime":"2025-11-25T19:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.101751 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.101793 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.101802 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.101819 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.101828 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.106727 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.106867 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.106732 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.106981 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.172011 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.172041 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.172050 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.172063 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.172072 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.194406 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:47Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.199394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.199538 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.199559 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.199593 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.199616 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.219663 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:47Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.225789 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.225831 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.225845 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.225867 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.225882 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.241225 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:47Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.252166 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.252245 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.252270 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.252302 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.252326 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.268735 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:47Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.274054 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.274132 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.274147 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.274173 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.274190 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.291400 4759 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fb282a89-ea25-4f14-b040-5cf9a95acf23\\\",\\\"systemUUID\\\":\\\"4ddc72e5-0e0b-4a87-ad4f-9874c7ee8269\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:47Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:47 crc kubenswrapper[4759]: E1125 19:34:47.291586 4759 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.293930 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.293986 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.294003 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.294025 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.294043 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.396315 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.396379 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.396399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.396424 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.396442 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.499090 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.499132 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.499140 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.499155 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.499165 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.602007 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.602062 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.602071 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.602090 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.602099 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.705533 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.705592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.705606 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.705625 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.705638 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.808486 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.808546 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.808563 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.808591 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.808611 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.911920 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.911981 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.912001 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.912026 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:47 crc kubenswrapper[4759]: I1125 19:34:47.912044 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:47Z","lastTransitionTime":"2025-11-25T19:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.014649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.014689 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.014700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.014717 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.014727 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.106510 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.106670 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:48 crc kubenswrapper[4759]: E1125 19:34:48.106869 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:48 crc kubenswrapper[4759]: E1125 19:34:48.107010 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.117649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.117710 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.117728 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.117751 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.117769 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.221850 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.221917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.221938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.221968 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.221989 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.325554 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.325602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.325615 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.325632 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.325644 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.429992 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.430072 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.430133 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.430173 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.430196 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.532815 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.532860 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.532871 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.532886 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.532896 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.635547 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.635649 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.635666 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.635719 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.635757 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.738410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.738482 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.738496 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.738512 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.738522 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.841284 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.841319 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.841333 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.841346 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.841355 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.943115 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.943157 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.943167 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.943182 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:48 crc kubenswrapper[4759]: I1125 19:34:48.943191 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:48Z","lastTransitionTime":"2025-11-25T19:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.045499 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.045543 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.045555 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.045572 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.045584 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.106781 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.106825 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:49 crc kubenswrapper[4759]: E1125 19:34:49.106957 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:49 crc kubenswrapper[4759]: E1125 19:34:49.107139 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.147821 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.147938 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.147960 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.147987 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.148005 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.250313 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.250353 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.250367 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.250384 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.250400 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.353357 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.353399 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.353410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.353429 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.353439 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.464394 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.464498 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.464523 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.464551 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.464572 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.567510 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.567591 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.567602 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.567617 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.567626 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.671382 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.671480 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.671537 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.671569 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.671596 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.774566 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.774623 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.774640 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.774664 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.774682 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.877386 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.877498 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.877517 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.877544 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.877562 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.980930 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.980997 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.981015 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.981039 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:49 crc kubenswrapper[4759]: I1125 19:34:49.981057 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:49Z","lastTransitionTime":"2025-11-25T19:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.083434 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.083509 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.083517 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.083559 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.083570 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.106614 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.106620 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:50 crc kubenswrapper[4759]: E1125 19:34:50.106716 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:50 crc kubenswrapper[4759]: E1125 19:34:50.106826 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.119486 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.131618 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.145294 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.158015 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.170933 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.181537 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.185373 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.185412 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.185424 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.185440 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.185473 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.191708 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.201736 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f6e9985-4676-472c-82bd-291bc8b3fa3a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d5d6e432bf167987c3da881c52aecd060161d9a794e0545a58bf2a31ee9ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.219239 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.231173 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.250094 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.265518 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.281399 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.288605 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.288780 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.288892 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.289024 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.289107 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.294934 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.317270 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.329617 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.340860 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.353305 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:50Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.391233 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.391296 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.391309 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.391329 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.391341 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.494099 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.494148 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.494163 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.494184 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.494200 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.597298 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.597344 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.597354 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.597373 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.597386 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.699655 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.699698 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.699719 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.699739 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.699754 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.802776 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.802832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.802848 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.802869 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.802885 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.906017 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.906060 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.906068 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.906082 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:50 crc kubenswrapper[4759]: I1125 19:34:50.906092 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:50Z","lastTransitionTime":"2025-11-25T19:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.009500 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.009576 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.009598 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.009634 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.009658 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.106987 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.107677 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.108301 4759 scope.go:117] "RemoveContainer" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" Nov 25 19:34:51 crc kubenswrapper[4759]: E1125 19:34:51.108814 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:51 crc kubenswrapper[4759]: E1125 19:34:51.109086 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.119867 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.119925 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.119998 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.120023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.120117 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.129952 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.222605 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.222650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.222667 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.222691 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.222707 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.325377 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.325425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.325443 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.325502 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.325522 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.427903 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.427942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.427950 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.427964 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.427973 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.530724 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.530766 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.530778 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.530796 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.530808 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.622975 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/2.log" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.625726 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.633391 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.633429 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.633459 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.633477 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.633487 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.638418 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f6e9985-4676-472c-82bd-291bc8b3fa3a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d5d6e432bf167987c3da881c52aecd060161d9a794e0545a58bf2a31ee9ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.652538 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.669151 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.682645 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.694261 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.707541 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.721063 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.735731 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.735771 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.735783 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.735801 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.735813 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.738757 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.749882 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.759973 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.770112 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.783379 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.794717 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.803757 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.822726 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea6d4cd-ccb1-4cf8-92c9-d39f05b39dca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df09a503ad4d85bb0383d070f52ac979188cbd361d3dc318ae1e96742daebf4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85f46c4fd1e2e9f85df12f3b4e03d903a1f2eae9ba61950682af19c45d0964aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee63675d5b4dac84d9cc76b94052791eae19e67b1b6ca63b5641bb45731df462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f20ac2001ecb232136b91658f51968bee70d612dd01f7be1708b228a686e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4aa15e3a239ee31d509b20eac95c20c7a636b6c7c2c80cf65f5b386ca07d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.835085 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.837858 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.838072 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.838216 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.838341 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.838531 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.847312 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.858478 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.870002 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.941069 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.941335 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.941483 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.941598 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:51 crc kubenswrapper[4759]: I1125 19:34:51.941688 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:51Z","lastTransitionTime":"2025-11-25T19:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.045109 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.045144 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.045153 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.045169 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.045180 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.106161 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:52 crc kubenswrapper[4759]: E1125 19:34:52.106273 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.106161 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:52 crc kubenswrapper[4759]: E1125 19:34:52.106371 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.147400 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.147478 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.147491 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.147513 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.147527 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.250387 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.250429 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.250469 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.250488 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.250501 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.352804 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.352842 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.352851 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.352882 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.352890 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.455696 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.455751 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.455763 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.455780 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.456232 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.558290 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.558334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.558345 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.558362 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.558376 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.629622 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/3.log" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.630499 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/2.log" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.632847 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" exitCode=1 Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.632884 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.632918 4759 scope.go:117] "RemoveContainer" containerID="40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.633934 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:34:52 crc kubenswrapper[4759]: E1125 19:34:52.634161 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.644540 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jz4pb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb894524-1c88-4838-aa0c-0a1956951d23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5ae2cca9a2ea4d40ab4c909aea0b600758cdbba63d1d5204e55048d9a8d825f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfmwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jz4pb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.654190 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07c1b558-ca9e-4edc-9420-05750ceee3a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8adb1ddb4089c039cc333c3c5180329067ad63bb976ff39f5683257912563b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sdj5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfsr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.660131 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.660177 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.660194 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.660215 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.660230 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.677157 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea6d4cd-ccb1-4cf8-92c9-d39f05b39dca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df09a503ad4d85bb0383d070f52ac979188cbd361d3dc318ae1e96742daebf4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85f46c4fd1e2e9f85df12f3b4e03d903a1f2eae9ba61950682af19c45d0964aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee63675d5b4dac84d9cc76b94052791eae19e67b1b6ca63b5641bb45731df462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f20ac2001ecb232136b91658f51968bee70d612dd01f7be1708b228a686e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4aa15e3a239ee31d509b20eac95c20c7a636b6c7c2c80cf65f5b386ca07d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f50ae40292bd8da45d08538579e1f57b5314e743294bd946d3ab4ee4c181d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a10fa353040c2bb09d72b30e7231ffdb64e36e80c190261e1c0fdb673b6bebdc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a0ed56cec87e34f9d2594548d0cdbb7ed72daa30ad74015d6a4950010e1af4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.693018 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9b1f292-6844-43a6-bb02-a0e661e3061e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T19:33:49Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 19:33:34.144567 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 19:33:34.145278 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1292959913/tls.crt::/tmp/serving-cert-1292959913/tls.key\\\\\\\"\\\\nI1125 19:33:49.495263 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 19:33:49.501944 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 19:33:49.501990 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 19:33:49.502033 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 19:33:49.502050 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 19:33:49.521013 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 19:33:49.521071 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521085 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 19:33:49.521099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 19:33:49.521107 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 19:33:49.521119 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 19:33:49.521128 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 19:33:49.521300 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 19:33:49.526231 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.709754 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.723602 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.737599 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9594e81bf805404d0ad4e200d8697243ee542b11844a6ffeec0f9cbe76c5476d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc4d5ff2f010ff55838bc14ff0b037884a0a42c6aab9d8bff1155934dcc79748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.749890 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://042624ff3cb0a2ade2944d6e8eccee114659cc1ae06e6650c69873a8aac7cbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.762980 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.763019 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.763032 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.763076 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.763088 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.764325 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vz6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67e7b861-9f72-41ba-ab98-35e0a2bdbe39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:37Z\\\",\\\"message\\\":\\\"2025-11-25T19:33:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279\\\\n2025-11-25T19:33:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cf4560ce-1a47-439c-b3a3-523bb1b5b279 to /host/opt/cni/bin/\\\\n2025-11-25T19:33:52Z [verbose] multus-daemon started\\\\n2025-11-25T19:33:52Z [verbose] Readiness Indicator file check\\\\n2025-11-25T19:34:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8q88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vz6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.776338 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e41227a-a128-4f1e-a21c-2242cea52970\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgmgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k5q6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.787679 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f6e9985-4676-472c-82bd-291bc8b3fa3a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d5d6e432bf167987c3da881c52aecd060161d9a794e0545a58bf2a31ee9ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2abb0ff513b85f423d440150042294774e5e4fb1c3a4ebd730aca677c8911acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.798990 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2115677-298c-4ccf-a72e-f9eb5669b16f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac99a26ab4599808b5a4096e38c27869ff59415e00d502ecb5f294dc85ace82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39d4e918efd857d9015e9968f578f655dacb8bfe8514a95720ed340db540a694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77861794288d59f6d0a9ce4d2dc6f0c259bba671ebe78ff0bff91fe7ef4a5425\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9773d85df131d129bd090c256c9829a580b140241a4c8a887565460f5258fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.808131 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ttvrw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3822015e-9df0-4cce-9d59-92247b44348b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2041d2789eb389e51b7028b4b0bf8e14e179be797eecbe7a30001bce9c3decf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qp56\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ttvrw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.842925 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-58fww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6dff2992-ff1a-4dce-b635-ee0f16ba11d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b21403da74c4010222d98756d318b42b6d0ecd9f04775c1a1af169fe6858af5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5efbdf644d2dfa8f55387ad230fefb803b7c6dc24b61334157506e83b63bc182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2056878a621184cdf05a637786c736e7c8fdb9ce49ef34c8f9518d964725426\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87f104da9634285700e2991fdb7b5d2f9c8915fc0b2624ed8a2f1ead28d55c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96f19aec40b9526d2639fcea755ab18529dd0d010486141bebb09144a28d1ed6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d26690d73c89622c77e825b168d52fb13e7aec2c19f8e2faf770da1f39be12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://259507bd825459a36ce3dbfb3eb3c7987389ef8e85b47ddfbd0784169863cf22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ndv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-58fww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.855190 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97fd7d2-78da-44c3-abea-d43bd0911f02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3418d7bea78ddf4c342ae4ac714ef5742fbc0a62e2b9d2eb1393f5ae25ed9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed03b58036b1b9b3e403fc0d6d0ea104447faf4132e8235ae86721ee4d5eaa75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f92f30d589c59005a3f01a32222af6afdf3cdfb2eda47be7dc808d29de3098e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.866216 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.866289 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.866305 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.866343 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.866356 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.867222 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1858b3a1bd21840d3da7e656c6d01b7b06d86552f1cfe53bd004c87febc0827f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.879333 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.897930 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b6037c-81e1-4fe8-9216-5b1863a029f1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:33:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40c0eb068338186197222cd3b92d7004f927113bda4e8145cbe34c08d14c35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:22Z\\\",\\\"message\\\":\\\" reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 19:34:21.994970 6420 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 19:34:21.995333 6420 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 19:34:21.996339 6420 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 19:34:21.996374 6420 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 19:34:21.996420 6420 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 19:34:21.996437 6420 factory.go:656] Stopping watch factory\\\\nI1125 19:34:21.996495 6420 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 19:34:22.004918 6420 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1125 19:34:22.004948 6420 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1125 19:34:22.005023 6420 ovnkube.go:599] Stopped ovnkube\\\\nI1125 19:34:22.005059 6420 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 19:34:22.005137 6420 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T19:34:52Z\\\",\\\"message\\\":\\\" 6821 services_controller.go:356] Processing sync for service openshift-kube-apiserver-operator/metrics for network=default\\\\nI1125 19:34:51.979896 6821 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nF1125 19:34:51.979898 6821 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:51Z is after 2025-08-24T17:21:41Z]\\\\nI1125 19:34:51.979783 6821 lb_config.go:1031] Cluster endpoints for openshift-console/downloads for network=default are: map[]\\\\nI1125 19:34:51.979917 6821 services_controller.go:443] Built service openshift-console/downlo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T19:34:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T19:33:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T19:33:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lr5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:33:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-99nmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.908908 4759 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd110cd-4afd-4c0a-b5a6-5a1062326603\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T19:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4f6eb7f5d478412f41e9350abf538cdbc41956e1e1a479cf5de98a10ae2ee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2db720de287a183bcd21f822f5b5a4ce9360386ab37c1c8f2f7b4141e9c52355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T19:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ltr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T19:34:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vczjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T19:34:52Z is after 2025-08-24T17:21:41Z" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.969023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.969052 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.969060 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.969073 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:52 crc kubenswrapper[4759]: I1125 19:34:52.969081 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:52Z","lastTransitionTime":"2025-11-25T19:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.071658 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.071904 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.071917 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.071931 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.071940 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.106412 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.106495 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:53 crc kubenswrapper[4759]: E1125 19:34:53.106580 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:53 crc kubenswrapper[4759]: E1125 19:34:53.106637 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.175632 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.175666 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.175674 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.175689 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.175700 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.277706 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.277751 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.277767 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.277797 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.277813 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.380439 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.380486 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.380494 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.380510 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.380520 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.482746 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.482782 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.482794 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.482810 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.482823 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.586376 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.586474 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.586493 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.586519 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.586538 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.637359 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/3.log" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.688690 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.688734 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.688742 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.688755 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.688763 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.791305 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.791356 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.791378 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.791408 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.791427 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.894677 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.894719 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.894731 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.894747 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.894762 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.996403 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.996470 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.996483 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.996501 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:53 crc kubenswrapper[4759]: I1125 19:34:53.996513 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:53Z","lastTransitionTime":"2025-11-25T19:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.098288 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.098332 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.098343 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.098360 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.098372 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.106579 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.106714 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.106854 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.106962 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.153093 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.153244 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153293 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:35:58.153262709 +0000 UTC m=+148.739210420 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.153352 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153372 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153394 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153407 4759 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153476 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 19:35:58.153442364 +0000 UTC m=+148.739390155 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.153523 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153530 4759 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153591 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:35:58.153571067 +0000 UTC m=+148.739518768 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.153618 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153707 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153721 4759 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153731 4759 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153766 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 19:35:58.153757072 +0000 UTC m=+148.739704773 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153812 4759 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: E1125 19:34:54.153840 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 19:35:58.153831904 +0000 UTC m=+148.739779725 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.200744 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.200801 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.200812 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.200829 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.200841 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.302747 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.302774 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.302782 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.302795 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.302803 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.405497 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.405584 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.405598 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.405613 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.405639 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.508151 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.508178 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.508185 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.508199 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.508207 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.610946 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.611074 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.611146 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.611241 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.611269 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.713906 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.714179 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.714204 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.714229 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.714247 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.816343 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.816379 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.816390 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.816404 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.816413 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.918404 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.918518 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.918542 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.918570 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:54 crc kubenswrapper[4759]: I1125 19:34:54.918597 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:54Z","lastTransitionTime":"2025-11-25T19:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.021244 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.021358 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.021381 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.021410 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.021432 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.106797 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:55 crc kubenswrapper[4759]: E1125 19:34:55.106900 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.106801 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:55 crc kubenswrapper[4759]: E1125 19:34:55.107178 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.123524 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.123562 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.123576 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.123592 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.123604 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.226170 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.226244 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.226268 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.226299 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.226322 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.328867 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.328896 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.328906 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.328936 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.328945 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.432250 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.432317 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.432334 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.432357 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.432373 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.535598 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.535679 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.535696 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.535715 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.535763 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.638734 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.638791 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.638808 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.638832 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.638850 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.741405 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.741473 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.741488 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.741504 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.741514 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.843609 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.843653 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.843664 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.843683 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.843697 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.946365 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.946417 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.946431 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.946469 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:55 crc kubenswrapper[4759]: I1125 19:34:55.946483 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:55Z","lastTransitionTime":"2025-11-25T19:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.048659 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.048716 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.048729 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.048768 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.048781 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.106707 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:56 crc kubenswrapper[4759]: E1125 19:34:56.106924 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.106728 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:56 crc kubenswrapper[4759]: E1125 19:34:56.107362 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.151763 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.151826 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.151853 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.151881 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.151902 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.254640 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.254707 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.254730 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.254757 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.254778 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.357701 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.357747 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.357760 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.357777 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.357788 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.460611 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.460644 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.460654 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.460668 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.460677 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.564030 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.564080 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.564093 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.564111 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.564123 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.666651 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.666700 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.666719 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.666740 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.666758 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.770379 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.770425 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.770435 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.770491 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.770505 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.873560 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.873605 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.873618 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.873633 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.873644 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.976556 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.976815 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.976909 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.976995 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:56 crc kubenswrapper[4759]: I1125 19:34:56.977061 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:56Z","lastTransitionTime":"2025-11-25T19:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.080309 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.080638 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.080705 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.080804 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.080894 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.105890 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.105911 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:57 crc kubenswrapper[4759]: E1125 19:34:57.106047 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:34:57 crc kubenswrapper[4759]: E1125 19:34:57.106167 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.183665 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.183908 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.183976 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.184050 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.184143 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.286736 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.286759 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.286767 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.286781 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.286790 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.389497 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.389536 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.389548 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.389564 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.389579 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.491601 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.491642 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.491654 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.491670 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.491681 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.586023 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.586090 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.586108 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.586131 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.586149 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.616872 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.616926 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.616942 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.616965 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.616982 4759 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T19:34:57Z","lastTransitionTime":"2025-11-25T19:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.645638 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4"] Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.646436 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.652988 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.653035 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.653327 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.653339 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.682938 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podStartSLOduration=67.682914899 podStartE2EDuration="1m7.682914899s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.682807856 +0000 UTC m=+88.268755567" watchObservedRunningTime="2025-11-25 19:34:57.682914899 +0000 UTC m=+88.268862610" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.683252 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-jz4pb" podStartSLOduration=67.683244927 podStartE2EDuration="1m7.683244927s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.666566229 +0000 UTC m=+88.252514020" watchObservedRunningTime="2025-11-25 19:34:57.683244927 +0000 UTC m=+88.269192648" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.729282 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vz6pj" podStartSLOduration=67.729255064 podStartE2EDuration="1m7.729255064s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.729074329 +0000 UTC m=+88.315022040" watchObservedRunningTime="2025-11-25 19:34:57.729255064 +0000 UTC m=+88.315202805" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.772436 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.7724182630000005 podStartE2EDuration="6.772418263s" podCreationTimestamp="2025-11-25 19:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.772250978 +0000 UTC m=+88.358198689" watchObservedRunningTime="2025-11-25 19:34:57.772418263 +0000 UTC m=+88.358365964" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.789426 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b260af0a-3afe-4fb9-9963-a307ffdb50aa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.789531 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.789565 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b260af0a-3afe-4fb9-9963-a307ffdb50aa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.789630 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.789675 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b260af0a-3afe-4fb9-9963-a307ffdb50aa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.819364 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.819340843 podStartE2EDuration="1m7.819340843s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.793819398 +0000 UTC m=+88.379767139" watchObservedRunningTime="2025-11-25 19:34:57.819340843 +0000 UTC m=+88.405288554" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890608 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b260af0a-3afe-4fb9-9963-a307ffdb50aa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890666 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b260af0a-3afe-4fb9-9963-a307ffdb50aa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890709 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890734 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b260af0a-3afe-4fb9-9963-a307ffdb50aa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890775 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890869 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.890913 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b260af0a-3afe-4fb9-9963-a307ffdb50aa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.892152 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b260af0a-3afe-4fb9-9963-a307ffdb50aa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.895996 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b260af0a-3afe-4fb9-9963-a307ffdb50aa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.898961 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=37.898949491 podStartE2EDuration="37.898949491s" podCreationTimestamp="2025-11-25 19:34:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.898941241 +0000 UTC m=+88.484888952" watchObservedRunningTime="2025-11-25 19:34:57.898949491 +0000 UTC m=+88.484897192" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.899234 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.899228658 podStartE2EDuration="17.899228658s" podCreationTimestamp="2025-11-25 19:34:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.88475427 +0000 UTC m=+88.470701981" watchObservedRunningTime="2025-11-25 19:34:57.899228658 +0000 UTC m=+88.485176359" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.905943 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b260af0a-3afe-4fb9-9963-a307ffdb50aa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g7rt4\" (UID: \"b260af0a-3afe-4fb9-9963-a307ffdb50aa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.929945 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ttvrw" podStartSLOduration=67.929927433 podStartE2EDuration="1m7.929927433s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.915991398 +0000 UTC m=+88.501939099" watchObservedRunningTime="2025-11-25 19:34:57.929927433 +0000 UTC m=+88.515875134" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.946025 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-58fww" podStartSLOduration=67.946004895 podStartE2EDuration="1m7.946004895s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.930982662 +0000 UTC m=+88.516930373" watchObservedRunningTime="2025-11-25 19:34:57.946004895 +0000 UTC m=+88.531952596" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.946380 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.946375805 podStartE2EDuration="1m4.946375805s" podCreationTimestamp="2025-11-25 19:33:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:57.946163809 +0000 UTC m=+88.532111510" watchObservedRunningTime="2025-11-25 19:34:57.946375805 +0000 UTC m=+88.532323506" Nov 25 19:34:57 crc kubenswrapper[4759]: I1125 19:34:57.982393 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" Nov 25 19:34:57 crc kubenswrapper[4759]: W1125 19:34:57.997578 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb260af0a_3afe_4fb9_9963_a307ffdb50aa.slice/crio-e3ac5f4f3a1fa2be47e2b87a6de91adc5f8f83d7e18e2ccb3bf1781620a5ace7 WatchSource:0}: Error finding container e3ac5f4f3a1fa2be47e2b87a6de91adc5f8f83d7e18e2ccb3bf1781620a5ace7: Status 404 returned error can't find the container with id e3ac5f4f3a1fa2be47e2b87a6de91adc5f8f83d7e18e2ccb3bf1781620a5ace7 Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.019919 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vczjw" podStartSLOduration=67.019899849 podStartE2EDuration="1m7.019899849s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:58.018666766 +0000 UTC m=+88.604614467" watchObservedRunningTime="2025-11-25 19:34:58.019899849 +0000 UTC m=+88.605847550" Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.106778 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:34:58 crc kubenswrapper[4759]: E1125 19:34:58.106873 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.106778 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:34:58 crc kubenswrapper[4759]: E1125 19:34:58.106959 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.660815 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" event={"ID":"b260af0a-3afe-4fb9-9963-a307ffdb50aa","Type":"ContainerStarted","Data":"e4743c5ed757346ba37879d01abe5002b0f6f94fdb216362212af9b97f90ad9f"} Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.660894 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" event={"ID":"b260af0a-3afe-4fb9-9963-a307ffdb50aa","Type":"ContainerStarted","Data":"e3ac5f4f3a1fa2be47e2b87a6de91adc5f8f83d7e18e2ccb3bf1781620a5ace7"} Nov 25 19:34:58 crc kubenswrapper[4759]: I1125 19:34:58.676549 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g7rt4" podStartSLOduration=68.676527046 podStartE2EDuration="1m8.676527046s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:34:58.675050766 +0000 UTC m=+89.260998537" watchObservedRunningTime="2025-11-25 19:34:58.676527046 +0000 UTC m=+89.262474757" Nov 25 19:34:59 crc kubenswrapper[4759]: I1125 19:34:59.106845 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:34:59 crc kubenswrapper[4759]: E1125 19:34:59.107198 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:34:59 crc kubenswrapper[4759]: I1125 19:34:59.106918 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:34:59 crc kubenswrapper[4759]: E1125 19:34:59.107387 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:00 crc kubenswrapper[4759]: I1125 19:35:00.106681 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:00 crc kubenswrapper[4759]: I1125 19:35:00.106767 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:00 crc kubenswrapper[4759]: E1125 19:35:00.109201 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:00 crc kubenswrapper[4759]: E1125 19:35:00.109428 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:01 crc kubenswrapper[4759]: I1125 19:35:01.106953 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:01 crc kubenswrapper[4759]: I1125 19:35:01.106987 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:01 crc kubenswrapper[4759]: E1125 19:35:01.107204 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:01 crc kubenswrapper[4759]: E1125 19:35:01.107317 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:02 crc kubenswrapper[4759]: I1125 19:35:02.106163 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:02 crc kubenswrapper[4759]: I1125 19:35:02.106179 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:02 crc kubenswrapper[4759]: E1125 19:35:02.106696 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:02 crc kubenswrapper[4759]: E1125 19:35:02.106782 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:03 crc kubenswrapper[4759]: I1125 19:35:03.106125 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:03 crc kubenswrapper[4759]: I1125 19:35:03.106187 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:03 crc kubenswrapper[4759]: E1125 19:35:03.106264 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:03 crc kubenswrapper[4759]: E1125 19:35:03.106360 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:04 crc kubenswrapper[4759]: I1125 19:35:04.106689 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:04 crc kubenswrapper[4759]: I1125 19:35:04.106754 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:04 crc kubenswrapper[4759]: E1125 19:35:04.106968 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:04 crc kubenswrapper[4759]: E1125 19:35:04.107055 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:04 crc kubenswrapper[4759]: I1125 19:35:04.107968 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:35:04 crc kubenswrapper[4759]: E1125 19:35:04.108160 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:35:05 crc kubenswrapper[4759]: I1125 19:35:05.105830 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:05 crc kubenswrapper[4759]: E1125 19:35:05.105955 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:05 crc kubenswrapper[4759]: I1125 19:35:05.106402 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:05 crc kubenswrapper[4759]: E1125 19:35:05.106827 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:06 crc kubenswrapper[4759]: I1125 19:35:06.106716 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:06 crc kubenswrapper[4759]: I1125 19:35:06.106795 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:06 crc kubenswrapper[4759]: E1125 19:35:06.106897 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:06 crc kubenswrapper[4759]: E1125 19:35:06.106979 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:07 crc kubenswrapper[4759]: I1125 19:35:07.106483 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:07 crc kubenswrapper[4759]: I1125 19:35:07.106497 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:07 crc kubenswrapper[4759]: E1125 19:35:07.106790 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:07 crc kubenswrapper[4759]: E1125 19:35:07.107440 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:07 crc kubenswrapper[4759]: I1125 19:35:07.857832 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:35:07 crc kubenswrapper[4759]: I1125 19:35:07.859117 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:35:07 crc kubenswrapper[4759]: E1125 19:35:07.859371 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:35:08 crc kubenswrapper[4759]: I1125 19:35:08.106690 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:08 crc kubenswrapper[4759]: I1125 19:35:08.106916 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:08 crc kubenswrapper[4759]: E1125 19:35:08.107115 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:08 crc kubenswrapper[4759]: E1125 19:35:08.107340 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:09 crc kubenswrapper[4759]: I1125 19:35:09.106202 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:09 crc kubenswrapper[4759]: I1125 19:35:09.106208 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:09 crc kubenswrapper[4759]: E1125 19:35:09.106525 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:09 crc kubenswrapper[4759]: E1125 19:35:09.106671 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:09 crc kubenswrapper[4759]: I1125 19:35:09.431251 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:09 crc kubenswrapper[4759]: E1125 19:35:09.431578 4759 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:35:09 crc kubenswrapper[4759]: E1125 19:35:09.431690 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs podName:8e41227a-a128-4f1e-a21c-2242cea52970 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:13.431661137 +0000 UTC m=+164.017608878 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs") pod "network-metrics-daemon-k5q6v" (UID: "8e41227a-a128-4f1e-a21c-2242cea52970") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 19:35:10 crc kubenswrapper[4759]: I1125 19:35:10.106796 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:10 crc kubenswrapper[4759]: I1125 19:35:10.108898 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:10 crc kubenswrapper[4759]: E1125 19:35:10.109035 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:10 crc kubenswrapper[4759]: E1125 19:35:10.109256 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:11 crc kubenswrapper[4759]: I1125 19:35:11.106659 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:11 crc kubenswrapper[4759]: I1125 19:35:11.106702 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:11 crc kubenswrapper[4759]: E1125 19:35:11.106854 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:11 crc kubenswrapper[4759]: E1125 19:35:11.107052 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:12 crc kubenswrapper[4759]: I1125 19:35:12.106238 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:12 crc kubenswrapper[4759]: I1125 19:35:12.106565 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:12 crc kubenswrapper[4759]: E1125 19:35:12.106747 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:12 crc kubenswrapper[4759]: E1125 19:35:12.106813 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:13 crc kubenswrapper[4759]: I1125 19:35:13.105771 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:13 crc kubenswrapper[4759]: E1125 19:35:13.105893 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:13 crc kubenswrapper[4759]: I1125 19:35:13.105771 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:13 crc kubenswrapper[4759]: E1125 19:35:13.105983 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:14 crc kubenswrapper[4759]: I1125 19:35:14.106520 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:14 crc kubenswrapper[4759]: I1125 19:35:14.106531 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:14 crc kubenswrapper[4759]: E1125 19:35:14.106715 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:14 crc kubenswrapper[4759]: E1125 19:35:14.106813 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:15 crc kubenswrapper[4759]: I1125 19:35:15.106070 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:15 crc kubenswrapper[4759]: I1125 19:35:15.106099 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:15 crc kubenswrapper[4759]: E1125 19:35:15.106194 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:15 crc kubenswrapper[4759]: E1125 19:35:15.106387 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:16 crc kubenswrapper[4759]: I1125 19:35:16.106155 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:16 crc kubenswrapper[4759]: I1125 19:35:16.106171 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:16 crc kubenswrapper[4759]: E1125 19:35:16.106399 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:16 crc kubenswrapper[4759]: E1125 19:35:16.106781 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:17 crc kubenswrapper[4759]: I1125 19:35:17.106118 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:17 crc kubenswrapper[4759]: I1125 19:35:17.106149 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:17 crc kubenswrapper[4759]: E1125 19:35:17.106257 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:17 crc kubenswrapper[4759]: E1125 19:35:17.106368 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:18 crc kubenswrapper[4759]: I1125 19:35:18.106499 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:18 crc kubenswrapper[4759]: I1125 19:35:18.106850 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:18 crc kubenswrapper[4759]: E1125 19:35:18.106944 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:18 crc kubenswrapper[4759]: E1125 19:35:18.107059 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:18 crc kubenswrapper[4759]: I1125 19:35:18.107271 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:35:18 crc kubenswrapper[4759]: E1125 19:35:18.107539 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:35:19 crc kubenswrapper[4759]: I1125 19:35:19.106151 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:19 crc kubenswrapper[4759]: I1125 19:35:19.106196 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:19 crc kubenswrapper[4759]: E1125 19:35:19.106286 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:19 crc kubenswrapper[4759]: E1125 19:35:19.106623 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:20 crc kubenswrapper[4759]: I1125 19:35:20.106665 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:20 crc kubenswrapper[4759]: I1125 19:35:20.106662 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:20 crc kubenswrapper[4759]: E1125 19:35:20.108693 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:20 crc kubenswrapper[4759]: E1125 19:35:20.108908 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:21 crc kubenswrapper[4759]: I1125 19:35:21.106332 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:21 crc kubenswrapper[4759]: I1125 19:35:21.106342 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:21 crc kubenswrapper[4759]: E1125 19:35:21.106525 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:21 crc kubenswrapper[4759]: E1125 19:35:21.106614 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:22 crc kubenswrapper[4759]: I1125 19:35:22.106328 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:22 crc kubenswrapper[4759]: I1125 19:35:22.106336 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:22 crc kubenswrapper[4759]: E1125 19:35:22.106545 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:22 crc kubenswrapper[4759]: E1125 19:35:22.106660 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:23 crc kubenswrapper[4759]: I1125 19:35:23.105969 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:23 crc kubenswrapper[4759]: E1125 19:35:23.106088 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:23 crc kubenswrapper[4759]: I1125 19:35:23.105975 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:23 crc kubenswrapper[4759]: E1125 19:35:23.106203 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.106651 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.106675 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:24 crc kubenswrapper[4759]: E1125 19:35:24.106781 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:24 crc kubenswrapper[4759]: E1125 19:35:24.106944 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.746043 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/1.log" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.746529 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/0.log" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.746568 4759 generic.go:334] "Generic (PLEG): container finished" podID="67e7b861-9f72-41ba-ab98-35e0a2bdbe39" containerID="224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353" exitCode=1 Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.746619 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerDied","Data":"224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353"} Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.746653 4759 scope.go:117] "RemoveContainer" containerID="105f0be5513d7aaaa1c32d04aa35b6b32884f13a7a677a9f2ea1b412fd0ecda6" Nov 25 19:35:24 crc kubenswrapper[4759]: I1125 19:35:24.747231 4759 scope.go:117] "RemoveContainer" containerID="224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353" Nov 25 19:35:24 crc kubenswrapper[4759]: E1125 19:35:24.747524 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vz6pj_openshift-multus(67e7b861-9f72-41ba-ab98-35e0a2bdbe39)\"" pod="openshift-multus/multus-vz6pj" podUID="67e7b861-9f72-41ba-ab98-35e0a2bdbe39" Nov 25 19:35:25 crc kubenswrapper[4759]: I1125 19:35:25.106320 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:25 crc kubenswrapper[4759]: E1125 19:35:25.106565 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:25 crc kubenswrapper[4759]: I1125 19:35:25.107396 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:25 crc kubenswrapper[4759]: E1125 19:35:25.107736 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:25 crc kubenswrapper[4759]: I1125 19:35:25.752120 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/1.log" Nov 25 19:35:26 crc kubenswrapper[4759]: I1125 19:35:26.106514 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:26 crc kubenswrapper[4759]: E1125 19:35:26.106989 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:26 crc kubenswrapper[4759]: I1125 19:35:26.106544 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:26 crc kubenswrapper[4759]: E1125 19:35:26.107720 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:27 crc kubenswrapper[4759]: I1125 19:35:27.106749 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:27 crc kubenswrapper[4759]: I1125 19:35:27.106764 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:27 crc kubenswrapper[4759]: E1125 19:35:27.106870 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:27 crc kubenswrapper[4759]: E1125 19:35:27.106965 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:28 crc kubenswrapper[4759]: I1125 19:35:28.106203 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:28 crc kubenswrapper[4759]: E1125 19:35:28.106379 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:28 crc kubenswrapper[4759]: I1125 19:35:28.106520 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:28 crc kubenswrapper[4759]: E1125 19:35:28.106819 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:29 crc kubenswrapper[4759]: I1125 19:35:29.106660 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:29 crc kubenswrapper[4759]: E1125 19:35:29.106795 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:29 crc kubenswrapper[4759]: I1125 19:35:29.106669 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:29 crc kubenswrapper[4759]: E1125 19:35:29.106887 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:30 crc kubenswrapper[4759]: E1125 19:35:30.086007 4759 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 19:35:30 crc kubenswrapper[4759]: I1125 19:35:30.106570 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:30 crc kubenswrapper[4759]: I1125 19:35:30.106737 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:30 crc kubenswrapper[4759]: E1125 19:35:30.108423 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:30 crc kubenswrapper[4759]: E1125 19:35:30.108654 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:30 crc kubenswrapper[4759]: I1125 19:35:30.109771 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:35:30 crc kubenswrapper[4759]: E1125 19:35:30.110016 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-99nmn_openshift-ovn-kubernetes(d8b6037c-81e1-4fe8-9216-5b1863a029f1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" Nov 25 19:35:30 crc kubenswrapper[4759]: E1125 19:35:30.209084 4759 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 19:35:31 crc kubenswrapper[4759]: I1125 19:35:31.106180 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:31 crc kubenswrapper[4759]: E1125 19:35:31.106337 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:31 crc kubenswrapper[4759]: I1125 19:35:31.106767 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:31 crc kubenswrapper[4759]: E1125 19:35:31.107755 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:32 crc kubenswrapper[4759]: I1125 19:35:32.106675 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:32 crc kubenswrapper[4759]: I1125 19:35:32.106831 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:32 crc kubenswrapper[4759]: E1125 19:35:32.107064 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:32 crc kubenswrapper[4759]: E1125 19:35:32.107389 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:33 crc kubenswrapper[4759]: I1125 19:35:33.106251 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:33 crc kubenswrapper[4759]: I1125 19:35:33.106298 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:33 crc kubenswrapper[4759]: E1125 19:35:33.106482 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:33 crc kubenswrapper[4759]: E1125 19:35:33.106588 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:34 crc kubenswrapper[4759]: I1125 19:35:34.106786 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:34 crc kubenswrapper[4759]: E1125 19:35:34.106891 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:34 crc kubenswrapper[4759]: I1125 19:35:34.107034 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:34 crc kubenswrapper[4759]: E1125 19:35:34.107096 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:35 crc kubenswrapper[4759]: I1125 19:35:35.106083 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:35 crc kubenswrapper[4759]: I1125 19:35:35.106142 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:35 crc kubenswrapper[4759]: E1125 19:35:35.106204 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:35 crc kubenswrapper[4759]: E1125 19:35:35.106266 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:35 crc kubenswrapper[4759]: E1125 19:35:35.210713 4759 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 19:35:36 crc kubenswrapper[4759]: I1125 19:35:36.106300 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:36 crc kubenswrapper[4759]: I1125 19:35:36.106312 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:36 crc kubenswrapper[4759]: E1125 19:35:36.106633 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:36 crc kubenswrapper[4759]: E1125 19:35:36.106830 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:37 crc kubenswrapper[4759]: I1125 19:35:37.106321 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:37 crc kubenswrapper[4759]: E1125 19:35:37.106549 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:37 crc kubenswrapper[4759]: I1125 19:35:37.106763 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:37 crc kubenswrapper[4759]: E1125 19:35:37.106933 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:38 crc kubenswrapper[4759]: I1125 19:35:38.106767 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:38 crc kubenswrapper[4759]: I1125 19:35:38.106899 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:38 crc kubenswrapper[4759]: E1125 19:35:38.106888 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:38 crc kubenswrapper[4759]: E1125 19:35:38.106952 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:38 crc kubenswrapper[4759]: I1125 19:35:38.107219 4759 scope.go:117] "RemoveContainer" containerID="224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353" Nov 25 19:35:38 crc kubenswrapper[4759]: I1125 19:35:38.800767 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/1.log" Nov 25 19:35:38 crc kubenswrapper[4759]: I1125 19:35:38.801427 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerStarted","Data":"f203afb0342346ff4202e818d51b048ac2643cdbdd1aeb099a142dc45aab802c"} Nov 25 19:35:39 crc kubenswrapper[4759]: I1125 19:35:39.106323 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:39 crc kubenswrapper[4759]: I1125 19:35:39.106361 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:39 crc kubenswrapper[4759]: E1125 19:35:39.106560 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:39 crc kubenswrapper[4759]: E1125 19:35:39.106680 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:40 crc kubenswrapper[4759]: I1125 19:35:40.106631 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:40 crc kubenswrapper[4759]: I1125 19:35:40.106710 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:40 crc kubenswrapper[4759]: E1125 19:35:40.108574 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:40 crc kubenswrapper[4759]: E1125 19:35:40.108776 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:40 crc kubenswrapper[4759]: E1125 19:35:40.211591 4759 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 19:35:41 crc kubenswrapper[4759]: I1125 19:35:41.106644 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:41 crc kubenswrapper[4759]: I1125 19:35:41.106639 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:41 crc kubenswrapper[4759]: E1125 19:35:41.106847 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:41 crc kubenswrapper[4759]: E1125 19:35:41.107019 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.106888 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.106951 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:42 crc kubenswrapper[4759]: E1125 19:35:42.107054 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:42 crc kubenswrapper[4759]: E1125 19:35:42.107305 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.108238 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.817758 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/3.log" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.820648 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerStarted","Data":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.821081 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.879881 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podStartSLOduration=112.879862374 podStartE2EDuration="1m52.879862374s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:35:42.879276779 +0000 UTC m=+133.465224470" watchObservedRunningTime="2025-11-25 19:35:42.879862374 +0000 UTC m=+133.465810095" Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.919545 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k5q6v"] Nov 25 19:35:42 crc kubenswrapper[4759]: I1125 19:35:42.919656 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:42 crc kubenswrapper[4759]: E1125 19:35:42.919760 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:43 crc kubenswrapper[4759]: I1125 19:35:43.106646 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:43 crc kubenswrapper[4759]: E1125 19:35:43.107120 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:44 crc kubenswrapper[4759]: I1125 19:35:44.106960 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:44 crc kubenswrapper[4759]: E1125 19:35:44.107149 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:44 crc kubenswrapper[4759]: I1125 19:35:44.107525 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:44 crc kubenswrapper[4759]: E1125 19:35:44.107690 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:45 crc kubenswrapper[4759]: I1125 19:35:45.106929 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:45 crc kubenswrapper[4759]: I1125 19:35:45.106980 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:45 crc kubenswrapper[4759]: E1125 19:35:45.107194 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:45 crc kubenswrapper[4759]: E1125 19:35:45.107262 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:45 crc kubenswrapper[4759]: E1125 19:35:45.213073 4759 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 19:35:46 crc kubenswrapper[4759]: I1125 19:35:46.106954 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:46 crc kubenswrapper[4759]: I1125 19:35:46.107022 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:46 crc kubenswrapper[4759]: E1125 19:35:46.107139 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:46 crc kubenswrapper[4759]: E1125 19:35:46.107257 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:47 crc kubenswrapper[4759]: I1125 19:35:47.106282 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:47 crc kubenswrapper[4759]: I1125 19:35:47.106313 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:47 crc kubenswrapper[4759]: E1125 19:35:47.106576 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:47 crc kubenswrapper[4759]: E1125 19:35:47.106820 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:48 crc kubenswrapper[4759]: I1125 19:35:48.106158 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:48 crc kubenswrapper[4759]: E1125 19:35:48.106335 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:48 crc kubenswrapper[4759]: I1125 19:35:48.106483 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:48 crc kubenswrapper[4759]: E1125 19:35:48.106668 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:49 crc kubenswrapper[4759]: I1125 19:35:49.106231 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:49 crc kubenswrapper[4759]: E1125 19:35:49.106433 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k5q6v" podUID="8e41227a-a128-4f1e-a21c-2242cea52970" Nov 25 19:35:49 crc kubenswrapper[4759]: I1125 19:35:49.107292 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:49 crc kubenswrapper[4759]: E1125 19:35:49.107702 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 19:35:50 crc kubenswrapper[4759]: I1125 19:35:50.106374 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:50 crc kubenswrapper[4759]: I1125 19:35:50.106506 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:50 crc kubenswrapper[4759]: E1125 19:35:50.107260 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 19:35:50 crc kubenswrapper[4759]: E1125 19:35:50.107469 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.106562 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.106592 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.108770 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.109312 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.109657 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 19:35:51 crc kubenswrapper[4759]: I1125 19:35:51.110154 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 19:35:52 crc kubenswrapper[4759]: I1125 19:35:52.106261 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:52 crc kubenswrapper[4759]: I1125 19:35:52.106348 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:52 crc kubenswrapper[4759]: I1125 19:35:52.110082 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 19:35:52 crc kubenswrapper[4759]: I1125 19:35:52.110342 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 19:35:54 crc kubenswrapper[4759]: I1125 19:35:54.896821 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:35:54 crc kubenswrapper[4759]: I1125 19:35:54.896890 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.253531 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.253643 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.253667 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.253708 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.253726 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:58 crc kubenswrapper[4759]: E1125 19:35:58.254241 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:38:00.254197111 +0000 UTC m=+270.840144802 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.255323 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.262185 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.262345 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.262628 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.348625 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.438301 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.449210 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:58 crc kubenswrapper[4759]: W1125 19:35:58.605381 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-7bb68bb2f54a7acf3f2998b8fec430420573fc6578a7a9ca5a63c832b1721372 WatchSource:0}: Error finding container 7bb68bb2f54a7acf3f2998b8fec430420573fc6578a7a9ca5a63c832b1721372: Status 404 returned error can't find the container with id 7bb68bb2f54a7acf3f2998b8fec430420573fc6578a7a9ca5a63c832b1721372 Nov 25 19:35:58 crc kubenswrapper[4759]: W1125 19:35:58.660940 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-585eb4b49b3999746b52594bc03eb67ea2b04321031814d26463a7bf31097306 WatchSource:0}: Error finding container 585eb4b49b3999746b52594bc03eb67ea2b04321031814d26463a7bf31097306: Status 404 returned error can't find the container with id 585eb4b49b3999746b52594bc03eb67ea2b04321031814d26463a7bf31097306 Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.878206 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b82758721dd1bccdfbabe9dfd4c82cd7d3be6b4c7c8df44946f869f2338d4e7b"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.878246 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a1bf708787ef3969884fd63cd2a45e5bc34670141eb4066c2555fef6a8ae69ed"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.879979 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d3442916fddb1586d69b76d0380c25240980cb658dc0bd1d8e2904423ac32657"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.880078 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"585eb4b49b3999746b52594bc03eb67ea2b04321031814d26463a7bf31097306"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.880226 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.882288 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d22d9a25ee227f44ed4519d5d4aac8b0546743dc983617cf1b3d85dbb4d8947e"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.882395 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7bb68bb2f54a7acf3f2998b8fec430420573fc6578a7a9ca5a63c832b1721372"} Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.905650 4759 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.954213 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.954627 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.957905 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.958232 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.958530 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959304 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ae46aa-b894-42d7-b294-f1185b120e6c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959345 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ae46aa-b894-42d7-b294-f1185b120e6c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959373 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959394 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959419 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959468 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z86ng\" (UniqueName: \"kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959509 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-auth-proxy-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959545 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eb03f454-a3ef-4b68-88e3-011d6a7cc832-machine-approver-tls\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959568 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959629 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqwq7\" (UniqueName: \"kubernetes.io/projected/f2ae46aa-b894-42d7-b294-f1185b120e6c-kube-api-access-kqwq7\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959652 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tbfs\" (UniqueName: \"kubernetes.io/projected/eb03f454-a3ef-4b68-88e3-011d6a7cc832-kube-api-access-2tbfs\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.959738 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.966654 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.966699 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.966779 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.966916 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.967096 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.967257 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.967495 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.967700 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.967893 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.968021 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.968136 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.968625 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mqv76"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.968685 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.968960 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.969192 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.976689 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qg7bm"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.977131 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.977976 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dnkhr"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.978163 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkz9g"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.978312 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.978378 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.978547 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.980600 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.980999 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.985729 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.986067 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.986754 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.986972 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.987005 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.987079 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.987134 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 19:35:58 crc kubenswrapper[4759]: I1125 19:35:58.987305 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:58.998749 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4szh4"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:58.999541 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.004889 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.010106 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-x8rv5"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.010596 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.011426 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.012583 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.013415 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.016698 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.016804 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.019002 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.019962 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.020487 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.020830 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.022767 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.022958 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.023193 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025119 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025256 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025277 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025646 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025736 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025815 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.025898 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.026004 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.026025 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vzt92"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.027723 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.027925 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.028065 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.028751 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.040224 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.042403 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.042735 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.042916 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043071 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043225 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043329 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043538 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043654 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043756 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.043936 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.044053 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.044261 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.045536 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.045748 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.046193 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.047262 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.047498 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.047535 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.051641 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.051890 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.052266 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.052501 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.053540 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.053603 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.053656 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.055839 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.058817 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.058986 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059079 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059335 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059549 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059637 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059716 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059789 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059810 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.059874 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060416 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060468 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z86ng\" (UniqueName: \"kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060502 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-auth-proxy-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060526 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eb03f454-a3ef-4b68-88e3-011d6a7cc832-machine-approver-tls\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060552 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060591 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tbfs\" (UniqueName: \"kubernetes.io/projected/eb03f454-a3ef-4b68-88e3-011d6a7cc832-kube-api-access-2tbfs\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060612 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqwq7\" (UniqueName: \"kubernetes.io/projected/f2ae46aa-b894-42d7-b294-f1185b120e6c-kube-api-access-kqwq7\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060633 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ae46aa-b894-42d7-b294-f1185b120e6c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060672 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ae46aa-b894-42d7-b294-f1185b120e6c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060693 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.060714 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.061314 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.061429 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.062203 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ae46aa-b894-42d7-b294-f1185b120e6c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.062436 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.062988 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.070661 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.071679 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.071970 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.073460 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.101203 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.102120 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eb03f454-a3ef-4b68-88e3-011d6a7cc832-auth-proxy-config\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.103059 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.103251 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ae46aa-b894-42d7-b294-f1185b120e6c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.103666 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eb03f454-a3ef-4b68-88e3-011d6a7cc832-machine-approver-tls\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.104586 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.106995 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z86ng\" (UniqueName: \"kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng\") pod \"route-controller-manager-6576b87f9c-xxqhr\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.117990 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqwq7\" (UniqueName: \"kubernetes.io/projected/f2ae46aa-b894-42d7-b294-f1185b120e6c-kube-api-access-kqwq7\") pod \"openshift-controller-manager-operator-756b6f6bc6-l98cg\" (UID: \"f2ae46aa-b894-42d7-b294-f1185b120e6c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.118493 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pvn5h"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.119882 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.120155 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tbfs\" (UniqueName: \"kubernetes.io/projected/eb03f454-a3ef-4b68-88e3-011d6a7cc832-kube-api-access-2tbfs\") pod \"machine-approver-56656f9798-p6xcd\" (UID: \"eb03f454-a3ef-4b68-88e3-011d6a7cc832\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.121175 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfwts"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.122350 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.125109 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.140987 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.141740 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.141775 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.141977 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142096 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142271 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142337 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142460 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142557 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142672 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142702 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142721 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142734 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142765 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142892 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142908 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142992 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143181 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143314 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mq5jf"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143339 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.142997 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143437 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143646 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143818 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143821 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.143926 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.144380 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.144457 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.144468 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.145227 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.147530 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tkwdq"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.148171 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.148966 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mqv76"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.148992 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.149090 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.149355 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.149893 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.151737 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.152320 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.154166 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-k7nmp"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.154667 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.155624 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.156943 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.159533 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.160132 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.162278 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.163186 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.163437 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164550 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-policies\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164593 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-config\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164625 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpll8\" (UniqueName: \"kubernetes.io/projected/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-kube-api-access-cpll8\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164684 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfmlf\" (UniqueName: \"kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164727 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dtc5\" (UniqueName: \"kubernetes.io/projected/124dfa3b-114d-4d88-961e-8b99f5530562-kube-api-access-2dtc5\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164760 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164830 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-dir\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164868 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-console-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164899 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbpfj\" (UniqueName: \"kubernetes.io/projected/65d156e7-cd25-4e40-a754-f9dc88fbf01f-kube-api-access-xbpfj\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164922 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-serving-cert\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164940 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-config\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164959 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-client\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.164989 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165019 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-client\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165042 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5229cef-43ed-49af-a71f-e389bc154e8d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165061 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165106 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c4b203-e670-4bca-9547-1cdbcaa58632-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165201 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z2dh\" (UniqueName: \"kubernetes.io/projected/54c4b203-e670-4bca-9547-1cdbcaa58632-kube-api-access-7z2dh\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165242 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-service-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165268 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-config\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165289 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124dfa3b-114d-4d88-961e-8b99f5530562-serving-cert\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165336 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c4b203-e670-4bca-9547-1cdbcaa58632-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165383 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65d156e7-cd25-4e40-a754-f9dc88fbf01f-serving-cert\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165423 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-encryption-config\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165466 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165489 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22pzs\" (UniqueName: \"kubernetes.io/projected/e5229cef-43ed-49af-a71f-e389bc154e8d-kube-api-access-22pzs\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165511 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzwc6\" (UniqueName: \"kubernetes.io/projected/4e9ac029-8117-49ff-be41-19f18e9f621a-kube-api-access-qzwc6\") pod \"downloads-7954f5f757-x8rv5\" (UID: \"4e9ac029-8117-49ff-be41-19f18e9f621a\") " pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165529 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-serving-cert\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165545 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165563 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-config\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165590 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8s24\" (UniqueName: \"kubernetes.io/projected/720df419-4c33-4e7f-9bfd-147191aa10e6-kube-api-access-w8s24\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165605 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165622 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-serving-cert\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165650 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-service-ca\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165694 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm45m\" (UniqueName: \"kubernetes.io/projected/32444b14-e189-4a50-8e25-ed6323de94c7-kube-api-access-pm45m\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165730 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165755 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-images\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165780 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68gsn\" (UniqueName: \"kubernetes.io/projected/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-kube-api-access-68gsn\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165807 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165832 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165852 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165876 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-oauth-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165920 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-oauth-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165945 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165967 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-trusted-ca-bundle\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.165991 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg8hx\" (UniqueName: \"kubernetes.io/projected/1bc8d591-5c1b-4703-9967-99f84f455cd4-kube-api-access-tg8hx\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.166013 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/124dfa3b-114d-4d88-961e-8b99f5530562-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.166035 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-trusted-ca\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.170631 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.171467 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.172409 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.175075 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.175700 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.180503 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.182295 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.182668 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wcq7p"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.182708 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.182972 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.183011 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.184660 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.186434 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.187207 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.187707 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.187972 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.190631 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.191597 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.192801 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.193884 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkz9g"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.195520 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.198176 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.198804 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.200044 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.202243 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4szh4"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.203788 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.205781 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mq5jf"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.207038 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wrlnm"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.210229 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.210761 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dnkhr"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.214568 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qg7bm"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.217089 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfwts"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.220249 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.220431 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.220951 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vzt92"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.222565 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pvn5h"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.225520 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-x8rv5"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.226683 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.228780 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.229884 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.231137 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.232037 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.234251 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.235531 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tkwdq"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.236968 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mqn2l"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.238127 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.238234 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.242844 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.246775 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wrlnm"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.249140 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.250032 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.250272 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.251595 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.252699 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mqn2l"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.253794 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.254848 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.256025 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.258475 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7zhk5"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.259248 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.259256 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7zhk5"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.263617 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267560 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47052dc0-053f-422a-814b-09d6d56f0a6d-metrics-tls\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267596 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z2dh\" (UniqueName: \"kubernetes.io/projected/54c4b203-e670-4bca-9547-1cdbcaa58632-kube-api-access-7z2dh\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267623 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-node-pullsecrets\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267641 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-service-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267657 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c4b203-e670-4bca-9547-1cdbcaa58632-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267675 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcqc6\" (UniqueName: \"kubernetes.io/projected/9e729de1-5c72-4804-8f28-67bb35e37b50-kube-api-access-lcqc6\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267751 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267793 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267822 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5fmd\" (UniqueName: \"kubernetes.io/projected/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-kube-api-access-d5fmd\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267845 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267874 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22pzs\" (UniqueName: \"kubernetes.io/projected/e5229cef-43ed-49af-a71f-e389bc154e8d-kube-api-access-22pzs\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267892 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267913 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62n84\" (UniqueName: \"kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267941 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv828\" (UniqueName: \"kubernetes.io/projected/540c9b5f-650e-4708-a665-ed3a7ea52b70-kube-api-access-kv828\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267970 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.267990 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-stats-auth\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268007 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268027 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mpj8\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-kube-api-access-8mpj8\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268046 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-serving-cert\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268065 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-service-ca\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268086 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm45m\" (UniqueName: \"kubernetes.io/projected/32444b14-e189-4a50-8e25-ed6323de94c7-kube-api-access-pm45m\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268104 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6k6h\" (UniqueName: \"kubernetes.io/projected/0c794333-1489-4d9a-b395-3648650ede06-kube-api-access-w6k6h\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268123 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268141 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268161 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35063c0f-1f00-4f84-9d90-986955cf006e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268180 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-images\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268200 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268219 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268239 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz689\" (UniqueName: \"kubernetes.io/projected/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-kube-api-access-nz689\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268258 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a44627-b436-4b88-bab8-0190ab7493e6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268278 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-oauth-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268294 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbljz\" (UniqueName: \"kubernetes.io/projected/f398f4ae-52d7-404e-94e8-84352c38f62a-kube-api-access-cbljz\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268316 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268335 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z24f\" (UniqueName: \"kubernetes.io/projected/a88a31d1-eda7-401b-b5e0-f14376043180-kube-api-access-5z24f\") pod \"migrator-59844c95c7-pb8fh\" (UID: \"a88a31d1-eda7-401b-b5e0-f14376043180\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268357 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-service-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268365 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-oauth-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268384 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.268419 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c4b203-e670-4bca-9547-1cdbcaa58632-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269341 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269604 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269659 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/124dfa3b-114d-4d88-961e-8b99f5530562-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269734 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-trusted-ca\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269762 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit-dir\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269781 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-srv-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269800 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce66674d-c2cc-4b29-87fb-33474d19da87-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269854 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg8hx\" (UniqueName: \"kubernetes.io/projected/1bc8d591-5c1b-4703-9967-99f84f455cd4-kube-api-access-tg8hx\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269873 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-cabundle\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269890 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269908 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269924 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msdhj\" (UniqueName: \"kubernetes.io/projected/3b1fec22-5e5d-46d8-82ae-03440a6df080-kube-api-access-msdhj\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269973 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfmlf\" (UniqueName: \"kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.269992 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270009 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-oauth-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270013 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270078 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c794333-1489-4d9a-b395-3648650ede06-tmpfs\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270114 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270148 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270191 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270207 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270235 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsknv\" (UniqueName: \"kubernetes.io/projected/48a44627-b436-4b88-bab8-0190ab7493e6-kube-api-access-wsknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270276 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270283 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270328 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270355 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270380 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cht75\" (UniqueName: \"kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270431 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-config\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270474 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-client\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270504 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270529 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270565 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/124dfa3b-114d-4d88-961e-8b99f5530562-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270611 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-client\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270640 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270666 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5229cef-43ed-49af-a71f-e389bc154e8d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270805 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.270875 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-service-ca\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271046 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-config\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271107 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124dfa3b-114d-4d88-961e-8b99f5530562-serving-cert\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271138 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljjzr\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-kube-api-access-ljjzr\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271195 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65d156e7-cd25-4e40-a754-f9dc88fbf01f-serving-cert\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271601 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-encryption-config\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271653 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271674 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271705 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271722 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-key\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271741 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271777 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-encryption-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271796 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj982\" (UniqueName: \"kubernetes.io/projected/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-kube-api-access-wj982\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271812 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64718f2c-6468-414f-a80f-a44ab3ec63a2-config\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271828 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271879 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzwc6\" (UniqueName: \"kubernetes.io/projected/4e9ac029-8117-49ff-be41-19f18e9f621a-kube-api-access-qzwc6\") pod \"downloads-7954f5f757-x8rv5\" (UID: \"4e9ac029-8117-49ff-be41-19f18e9f621a\") " pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271915 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271934 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-config\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271953 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8s24\" (UniqueName: \"kubernetes.io/projected/720df419-4c33-4e7f-9bfd-147191aa10e6-kube-api-access-w8s24\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.271998 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-serving-cert\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272020 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce66674d-c2cc-4b29-87fb-33474d19da87-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272174 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65d156e7-cd25-4e40-a754-f9dc88fbf01f-config\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272555 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-ca\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272609 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-serving-cert\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272628 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-profile-collector-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272657 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272679 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68gsn\" (UniqueName: \"kubernetes.io/projected/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-kube-api-access-68gsn\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272718 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272721 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-config\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272736 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272754 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.272840 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-serving-cert\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273182 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273582 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-trusted-ca\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273716 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-config\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273731 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9e729de1-5c72-4804-8f28-67bb35e37b50-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273814 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a44627-b436-4b88-bab8-0190ab7493e6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273857 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-apiservice-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273857 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273910 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124dfa3b-114d-4d88-961e-8b99f5530562-serving-cert\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273929 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-webhook-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.273961 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274015 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274047 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-encryption-config\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274050 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht4nl\" (UniqueName: \"kubernetes.io/projected/47052dc0-053f-422a-814b-09d6d56f0a6d-kube-api-access-ht4nl\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274186 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-oauth-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274303 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-images\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274315 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-trusted-ca-bundle\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274394 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-client\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274504 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-serving-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274824 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274960 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-serving-cert\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275027 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-trusted-ca-bundle\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.274965 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpll8\" (UniqueName: \"kubernetes.io/projected/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-kube-api-access-cpll8\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275077 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-image-import-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275094 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce66674d-c2cc-4b29-87fb-33474d19da87-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275133 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-policies\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275158 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275152 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-config\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275209 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275260 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dtc5\" (UniqueName: \"kubernetes.io/projected/124dfa3b-114d-4d88-961e-8b99f5530562-kube-api-access-2dtc5\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275400 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-service-ca-bundle\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275610 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5229cef-43ed-49af-a71f-e389bc154e8d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275623 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc8d591-5c1b-4703-9967-99f84f455cd4-config\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275633 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275656 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-policies\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275697 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn8c9\" (UniqueName: \"kubernetes.io/projected/64718f2c-6468-414f-a80f-a44ab3ec63a2-kube-api-access-xn8c9\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275721 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-dir\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275736 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-console-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275753 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-serving-cert\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275769 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbpfj\" (UniqueName: \"kubernetes.io/projected/65d156e7-cd25-4e40-a754-f9dc88fbf01f-kube-api-access-xbpfj\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275788 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275806 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6fgw\" (UniqueName: \"kubernetes.io/projected/ef151ed2-118a-4cb8-b197-ae81db91ef3c-kube-api-access-d6fgw\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275843 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c4b203-e670-4bca-9547-1cdbcaa58632-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275862 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-client\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275879 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275897 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64718f2c-6468-414f-a80f-a44ab3ec63a2-serving-cert\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.275914 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.276024 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/720df419-4c33-4e7f-9bfd-147191aa10e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.276298 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/720df419-4c33-4e7f-9bfd-147191aa10e6-audit-dir\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.276703 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32444b14-e189-4a50-8e25-ed6323de94c7-console-config\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.276883 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.277487 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.278052 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c4b203-e670-4bca-9547-1cdbcaa58632-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.278295 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-serving-cert\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.278638 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65d156e7-cd25-4e40-a754-f9dc88fbf01f-serving-cert\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.278983 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32444b14-e189-4a50-8e25-ed6323de94c7-console-serving-cert\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.279054 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bc8d591-5c1b-4703-9967-99f84f455cd4-etcd-client\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.281952 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.300000 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.313836 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.323350 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.331218 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.341095 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.341607 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.361159 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377863 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377903 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62n84\" (UniqueName: \"kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377922 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-stats-auth\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377943 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv828\" (UniqueName: \"kubernetes.io/projected/540c9b5f-650e-4708-a665-ed3a7ea52b70-kube-api-access-kv828\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377964 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-csi-data-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377981 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.377997 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378014 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378047 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378063 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mpj8\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-kube-api-access-8mpj8\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378094 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6k6h\" (UniqueName: \"kubernetes.io/projected/0c794333-1489-4d9a-b395-3648650ede06-kube-api-access-w6k6h\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378214 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378237 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35063c0f-1f00-4f84-9d90-986955cf006e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378253 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378268 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378283 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz689\" (UniqueName: \"kubernetes.io/projected/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-kube-api-access-nz689\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378308 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbljz\" (UniqueName: \"kubernetes.io/projected/f398f4ae-52d7-404e-94e8-84352c38f62a-kube-api-access-cbljz\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378323 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a44627-b436-4b88-bab8-0190ab7493e6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378352 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378394 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z24f\" (UniqueName: \"kubernetes.io/projected/a88a31d1-eda7-401b-b5e0-f14376043180-kube-api-access-5z24f\") pod \"migrator-59844c95c7-pb8fh\" (UID: \"a88a31d1-eda7-401b-b5e0-f14376043180\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378464 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378488 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378506 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce66674d-c2cc-4b29-87fb-33474d19da87-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378532 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit-dir\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378546 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-srv-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378563 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-cabundle\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378581 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378599 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378617 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msdhj\" (UniqueName: \"kubernetes.io/projected/3b1fec22-5e5d-46d8-82ae-03440a6df080-kube-api-access-msdhj\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378636 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378653 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378678 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378695 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c794333-1489-4d9a-b395-3648650ede06-tmpfs\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378711 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378728 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378747 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378764 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378776 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379107 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit-dir\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.378782 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsknv\" (UniqueName: \"kubernetes.io/projected/48a44627-b436-4b88-bab8-0190ab7493e6-kube-api-access-wsknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379249 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379278 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379304 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cht75\" (UniqueName: \"kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379334 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379361 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-registration-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379402 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379430 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379488 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljjzr\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-kube-api-access-ljjzr\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379514 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xhjh\" (UniqueName: \"kubernetes.io/projected/da4f223e-5556-440d-93ad-eacfc180ae21-kube-api-access-8xhjh\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379558 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzcjk\" (UniqueName: \"kubernetes.io/projected/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-kube-api-access-xzcjk\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379584 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379606 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379629 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379654 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379672 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-key\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379695 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h26sc\" (UniqueName: \"kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379715 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379746 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-encryption-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379770 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj982\" (UniqueName: \"kubernetes.io/projected/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-kube-api-access-wj982\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.379363 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c794333-1489-4d9a-b395-3648650ede06-tmpfs\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380000 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380003 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64718f2c-6468-414f-a80f-a44ab3ec63a2-config\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380039 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380064 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7c6d\" (UniqueName: \"kubernetes.io/projected/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-kube-api-access-r7c6d\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380083 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce66674d-c2cc-4b29-87fb-33474d19da87-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380100 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hwk7\" (UniqueName: \"kubernetes.io/projected/6bcc761d-f5bd-476b-a8e5-25dcca68942e-kube-api-access-4hwk7\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380120 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-serving-cert\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380139 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380154 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-profile-collector-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380189 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9e729de1-5c72-4804-8f28-67bb35e37b50-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380206 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-apiservice-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380223 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a44627-b436-4b88-bab8-0190ab7493e6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380243 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380259 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-socket-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380286 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-webhook-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380308 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380326 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380342 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht4nl\" (UniqueName: \"kubernetes.io/projected/47052dc0-053f-422a-814b-09d6d56f0a6d-kube-api-access-ht4nl\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380358 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-serving-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380376 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380398 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-image-import-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380415 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce66674d-c2cc-4b29-87fb-33474d19da87-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380438 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380473 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-service-ca-bundle\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380489 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-plugins-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380507 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn8c9\" (UniqueName: \"kubernetes.io/projected/64718f2c-6468-414f-a80f-a44ab3ec63a2-kube-api-access-xn8c9\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380542 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380563 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-audit\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380568 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6fgw\" (UniqueName: \"kubernetes.io/projected/ef151ed2-118a-4cb8-b197-ae81db91ef3c-kube-api-access-d6fgw\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380607 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-client\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380630 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380647 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64718f2c-6468-414f-a80f-a44ab3ec63a2-serving-cert\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380669 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-node-pullsecrets\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380684 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47052dc0-053f-422a-814b-09d6d56f0a6d-metrics-tls\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380709 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-mountpoint-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380728 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcqc6\" (UniqueName: \"kubernetes.io/projected/9e729de1-5c72-4804-8f28-67bb35e37b50-kube-api-access-lcqc6\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380750 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380768 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5fmd\" (UniqueName: \"kubernetes.io/projected/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-kube-api-access-d5fmd\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380786 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.380908 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.381230 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.381799 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-node-pullsecrets\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.382012 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.382535 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.384159 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.384161 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-serving-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.384600 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.385021 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-image-import-ca\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.385408 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.386110 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.386522 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.386877 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.389058 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.389436 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.389865 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.390507 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-etcd-client\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.391459 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.392475 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-encryption-config\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.393771 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-serving-cert\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.401188 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.406970 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.415963 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47052dc0-053f-422a-814b-09d6d56f0a6d-metrics-tls\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.420359 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.440530 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.461316 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482036 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-registration-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482085 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xhjh\" (UniqueName: \"kubernetes.io/projected/da4f223e-5556-440d-93ad-eacfc180ae21-kube-api-access-8xhjh\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482112 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzcjk\" (UniqueName: \"kubernetes.io/projected/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-kube-api-access-xzcjk\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482148 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h26sc\" (UniqueName: \"kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482187 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482341 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-registration-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482048 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.482209 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7c6d\" (UniqueName: \"kubernetes.io/projected/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-kube-api-access-r7c6d\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483003 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hwk7\" (UniqueName: \"kubernetes.io/projected/6bcc761d-f5bd-476b-a8e5-25dcca68942e-kube-api-access-4hwk7\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483070 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483091 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-socket-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483135 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-plugins-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483192 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-mountpoint-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483254 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-csi-data-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483278 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483299 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483320 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483746 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-socket-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483752 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-plugins-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483818 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-mountpoint-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.483896 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6bcc761d-f5bd-476b-a8e5-25dcca68942e-csi-data-dir\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.486461 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.486529 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.487274 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.513275 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a44627-b436-4b88-bab8-0190ab7493e6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.520561 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.521275 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.523643 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.534667 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.541855 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.546814 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg"] Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.555312 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.580111 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.585220 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a44627-b436-4b88-bab8-0190ab7493e6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.604054 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.621363 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.640219 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.660511 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.672099 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce66674d-c2cc-4b29-87fb-33474d19da87-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.680682 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.701263 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.711701 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce66674d-c2cc-4b29-87fb-33474d19da87-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.721793 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.740645 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.761145 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.786645 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.796145 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-key\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.800938 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.810402 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f398f4ae-52d7-404e-94e8-84352c38f62a-signing-cabundle\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.820494 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.840944 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.860701 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.880820 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.885864 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" event={"ID":"6f9089f5-e196-48ad-9d8c-a14544b6a96f","Type":"ContainerStarted","Data":"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.885907 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" event={"ID":"6f9089f5-e196-48ad-9d8c-a14544b6a96f","Type":"ContainerStarted","Data":"b8aaf6770879642414266e82093be4b9699f5c3adcdbf83d7e2b5f75414a93c6"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.886064 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.886636 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-webhook-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.886845 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c794333-1489-4d9a-b395-3648650ede06-apiservice-cert\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.886928 4759 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xxqhr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.887057 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.887540 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" event={"ID":"eb03f454-a3ef-4b68-88e3-011d6a7cc832","Type":"ContainerStarted","Data":"df160ccada06fc018417a54c597844818b09f2bdf4b952489100121d13d6274c"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.887596 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" event={"ID":"eb03f454-a3ef-4b68-88e3-011d6a7cc832","Type":"ContainerStarted","Data":"d8e38608372034ac4aadd2f07fb97e205e98d8a834cd7cf30ed82e8e9810cb9f"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.889259 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" event={"ID":"f2ae46aa-b894-42d7-b294-f1185b120e6c","Type":"ContainerStarted","Data":"fadd3ad6245941f66f937a7727eab768feb64cfe163bdb47213b01f60c9ef895"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.889283 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" event={"ID":"f2ae46aa-b894-42d7-b294-f1185b120e6c","Type":"ContainerStarted","Data":"16429bb2e3734fb570dae70f5532de481cd3279fa89c8098c44ab432b121b43f"} Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.900938 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.920552 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.943713 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.963276 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.980816 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 19:35:59 crc kubenswrapper[4759]: I1125 19:35:59.985546 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.000926 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.027108 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.031536 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.041100 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.061737 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.080758 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.092331 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.100903 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.120679 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.125561 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9e729de1-5c72-4804-8f28-67bb35e37b50-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.140973 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.159293 4759 request.go:700] Waited for 1.009207901s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.161597 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.166366 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-profile-collector-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.166737 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.166883 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.181839 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.192483 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3b1fec22-5e5d-46d8-82ae-03440a6df080-srv-cert\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.200897 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.220641 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.242038 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.256470 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64718f2c-6468-414f-a80f-a44ab3ec63a2-serving-cert\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.261039 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.271511 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64718f2c-6468-414f-a80f-a44ab3ec63a2-config\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.281087 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.301890 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.320167 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.348378 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.360862 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.376060 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-stats-auth\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379591 4759 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379670 4759 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379700 4759 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379733 4759 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379788 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images podName:ef151ed2-118a-4cb8-b197-ae81db91ef3c nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879646789 +0000 UTC m=+151.465594490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images") pod "machine-config-operator-74547568cd-lkwxn" (UID: "ef151ed2-118a-4cb8-b197-ae81db91ef3c") : failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379745 4759 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379839 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls podName:ef151ed2-118a-4cb8-b197-ae81db91ef3c nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879810924 +0000 UTC m=+151.465758625 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls") pod "machine-config-operator-74547568cd-lkwxn" (UID: "ef151ed2-118a-4cb8-b197-ae81db91ef3c") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379859 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert podName:35063c0f-1f00-4f84-9d90-986955cf006e nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879851845 +0000 UTC m=+151.465799546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert") pod "kube-controller-manager-operator-78b949d7b-9xhm8" (UID: "35063c0f-1f00-4f84-9d90-986955cf006e") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379875 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert podName:540c9b5f-650e-4708-a665-ed3a7ea52b70 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879868225 +0000 UTC m=+151.465815926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert") pod "olm-operator-6b444d44fb-fcwxs" (UID: "540c9b5f-650e-4708-a665-ed3a7ea52b70") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379919 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config podName:35063c0f-1f00-4f84-9d90-986955cf006e nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879907966 +0000 UTC m=+151.465855677 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config") pod "kube-controller-manager-operator-78b949d7b-9xhm8" (UID: "35063c0f-1f00-4f84-9d90-986955cf006e") : failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379773 4759 configmap.go:193] Couldn't get configMap openshift-ingress-operator/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.379970 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca podName:22d364b4-96d8-46fa-b66f-57eabab1d389 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.879951328 +0000 UTC m=+151.465899029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca") pod "ingress-operator-5b745b69d9-qqj26" (UID: "22d364b4-96d8-46fa-b66f-57eabab1d389") : failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.380852 4759 secret.go:188] Couldn't get secret openshift-ingress/router-metrics-certs-default: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.380897 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs podName:2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.880885903 +0000 UTC m=+151.466833654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs") pod "router-default-5444994796-k7nmp" (UID: "2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.381271 4759 secret.go:188] Couldn't get secret openshift-ingress-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.381314 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls podName:22d364b4-96d8-46fa-b66f-57eabab1d389 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.881303854 +0000 UTC m=+151.467251645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls") pod "ingress-operator-5b745b69d9-qqj26" (UID: "22d364b4-96d8-46fa-b66f-57eabab1d389") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.381595 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.384044 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-service-ca-bundle\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.384145 4759 secret.go:188] Couldn't get secret openshift-ingress/router-certs-default: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.384197 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate podName:2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.884184302 +0000 UTC m=+151.470132073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-certificate" (UniqueName: "kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate") pod "router-default-5444994796-k7nmp" (UID: "2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.400206 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.421101 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.442051 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.460060 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.480561 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484605 4759 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484688 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume podName:b9999873-da0d-4ed3-88ac-cdfd8865cdd2 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.984662639 +0000 UTC m=+151.570610340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume") pod "dns-default-mqn2l" (UID: "b9999873-da0d-4ed3-88ac-cdfd8865cdd2") : failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484873 4759 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484911 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls podName:b9999873-da0d-4ed3-88ac-cdfd8865cdd2 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.984904545 +0000 UTC m=+151.570852236 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls") pod "dns-default-mqn2l" (UID: "b9999873-da0d-4ed3-88ac-cdfd8865cdd2") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484954 4759 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.484985 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls podName:1e8f64c9-2b3b-4e0f-93ea-78f589d591c5 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.984975537 +0000 UTC m=+151.570923238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls") pod "machine-config-controller-84d6567774-6hvlv" (UID: "1e8f64c9-2b3b-4e0f-93ea-78f589d591c5") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.485028 4759 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.485145 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert podName:da4f223e-5556-440d-93ad-eacfc180ae21 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.985116281 +0000 UTC m=+151.571063982 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert") pod "ingress-canary-7zhk5" (UID: "da4f223e-5556-440d-93ad-eacfc180ae21") : failed to sync secret cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.487198 4759 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: E1125 19:36:00.487254 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume podName:a0289180-1fc0-4d66-86ce-4e09fee7e0a9 nodeName:}" failed. No retries permitted until 2025-11-25 19:36:00.987241909 +0000 UTC m=+151.573189610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume") pod "collect-profiles-29401650-lfdkx" (UID: "a0289180-1fc0-4d66-86ce-4e09fee7e0a9") : failed to sync configmap cache: timed out waiting for the condition Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.500701 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.521470 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.541061 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.560161 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.580745 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.601017 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.620823 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.640778 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.665585 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.680499 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.700576 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.720550 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.741128 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.761036 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.781205 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.802357 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.821741 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.841857 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.860162 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.880667 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.894162 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" event={"ID":"eb03f454-a3ef-4b68-88e3-011d6a7cc832","Type":"ContainerStarted","Data":"bb5f710de87c8a7f7c8ba4a38e98ae930b20e86bd5af81463c31971366efc878"} Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.898562 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.900648 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.911916 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.911991 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.912028 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.912053 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.912128 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.912999 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef151ed2-118a-4cb8-b197-ae81db91ef3c-images\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913073 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35063c0f-1f00-4f84-9d90-986955cf006e-config\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913118 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913210 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913321 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913432 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22d364b4-96d8-46fa-b66f-57eabab1d389-trusted-ca\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.913483 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.916469 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef151ed2-118a-4cb8-b197-ae81db91ef3c-proxy-tls\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.916777 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22d364b4-96d8-46fa-b66f-57eabab1d389-metrics-tls\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.917340 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-metrics-certs\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.917914 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35063c0f-1f00-4f84-9d90-986955cf006e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.918383 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/540c9b5f-650e-4708-a665-ed3a7ea52b70-srv-cert\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.919938 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-default-certificate\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.924326 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.940494 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.960544 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 19:36:00 crc kubenswrapper[4759]: I1125 19:36:00.981122 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.000680 4759 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.014372 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.014666 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.014861 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.015000 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.015251 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.015897 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.018802 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-proxy-tls\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.021227 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.040816 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.061205 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.080502 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.085987 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-config-volume\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.100186 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.108488 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-metrics-tls\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.121548 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.128238 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da4f223e-5556-440d-93ad-eacfc180ae21-cert\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.140992 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.160241 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.178761 4759 request.go:700] Waited for 1.919208362s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.179829 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.213728 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z2dh\" (UniqueName: \"kubernetes.io/projected/54c4b203-e670-4bca-9547-1cdbcaa58632-kube-api-access-7z2dh\") pod \"openshift-apiserver-operator-796bbdcf4f-rnw7l\" (UID: \"54c4b203-e670-4bca-9547-1cdbcaa58632\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.233856 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm45m\" (UniqueName: \"kubernetes.io/projected/32444b14-e189-4a50-8e25-ed6323de94c7-kube-api-access-pm45m\") pod \"console-f9d7485db-4szh4\" (UID: \"32444b14-e189-4a50-8e25-ed6323de94c7\") " pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.258506 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22pzs\" (UniqueName: \"kubernetes.io/projected/e5229cef-43ed-49af-a71f-e389bc154e8d-kube-api-access-22pzs\") pod \"cluster-samples-operator-665b6dd947-b9tgz\" (UID: \"e5229cef-43ed-49af-a71f-e389bc154e8d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.276789 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg8hx\" (UniqueName: \"kubernetes.io/projected/1bc8d591-5c1b-4703-9967-99f84f455cd4-kube-api-access-tg8hx\") pod \"etcd-operator-b45778765-dnkhr\" (UID: \"1bc8d591-5c1b-4703-9967-99f84f455cd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.287007 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.298368 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfmlf\" (UniqueName: \"kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf\") pod \"controller-manager-879f6c89f-ddq5p\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.319907 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8s24\" (UniqueName: \"kubernetes.io/projected/720df419-4c33-4e7f-9bfd-147191aa10e6-kube-api-access-w8s24\") pod \"apiserver-7bbb656c7d-r4wvg\" (UID: \"720df419-4c33-4e7f-9bfd-147191aa10e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.340931 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzwc6\" (UniqueName: \"kubernetes.io/projected/4e9ac029-8117-49ff-be41-19f18e9f621a-kube-api-access-qzwc6\") pod \"downloads-7954f5f757-x8rv5\" (UID: \"4e9ac029-8117-49ff-be41-19f18e9f621a\") " pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.356765 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68gsn\" (UniqueName: \"kubernetes.io/projected/d74b7f2c-ab7c-4618-a700-5b43b5e65cb6-kube-api-access-68gsn\") pod \"console-operator-58897d9998-qg7bm\" (UID: \"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6\") " pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.376911 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dtc5\" (UniqueName: \"kubernetes.io/projected/124dfa3b-114d-4d88-961e-8b99f5530562-kube-api-access-2dtc5\") pod \"openshift-config-operator-7777fb866f-ngvqb\" (UID: \"124dfa3b-114d-4d88-961e-8b99f5530562\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.399034 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbpfj\" (UniqueName: \"kubernetes.io/projected/65d156e7-cd25-4e40-a754-f9dc88fbf01f-kube-api-access-xbpfj\") pod \"authentication-operator-69f744f599-mkz9g\" (UID: \"65d156e7-cd25-4e40-a754-f9dc88fbf01f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.416728 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpll8\" (UniqueName: \"kubernetes.io/projected/0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16-kube-api-access-cpll8\") pod \"machine-api-operator-5694c8668f-mqv76\" (UID: \"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.451513 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.457496 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62n84\" (UniqueName: \"kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84\") pod \"oauth-openshift-558db77b4-pvn5h\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.460743 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.472033 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.478258 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv828\" (UniqueName: \"kubernetes.io/projected/540c9b5f-650e-4708-a665-ed3a7ea52b70-kube-api-access-kv828\") pod \"olm-operator-6b444d44fb-fcwxs\" (UID: \"540c9b5f-650e-4708-a665-ed3a7ea52b70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.499548 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mpj8\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-kube-api-access-8mpj8\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.510782 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.517690 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6k6h\" (UniqueName: \"kubernetes.io/projected/0c794333-1489-4d9a-b395-3648650ede06-kube-api-access-w6k6h\") pod \"packageserver-d55dfcdfc-w5jkg\" (UID: \"0c794333-1489-4d9a-b395-3648650ede06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.534746 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4szh4"] Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.545462 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.548927 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35063c0f-1f00-4f84-9d90-986955cf006e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9xhm8\" (UID: \"35063c0f-1f00-4f84-9d90-986955cf006e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.550887 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.553719 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.558167 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsknv\" (UniqueName: \"kubernetes.io/projected/48a44627-b436-4b88-bab8-0190ab7493e6-kube-api-access-wsknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8bf6n\" (UID: \"48a44627-b436-4b88-bab8-0190ab7493e6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.569930 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.582663 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z24f\" (UniqueName: \"kubernetes.io/projected/a88a31d1-eda7-401b-b5e0-f14376043180-kube-api-access-5z24f\") pod \"migrator-59844c95c7-pb8fh\" (UID: \"a88a31d1-eda7-401b-b5e0-f14376043180\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.589768 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.597131 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz689\" (UniqueName: \"kubernetes.io/projected/32fdca48-e90c-4f5e-8524-3abaeb2f1bfb-kube-api-access-nz689\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mlsn\" (UID: \"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.606981 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.615468 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msdhj\" (UniqueName: \"kubernetes.io/projected/3b1fec22-5e5d-46d8-82ae-03440a6df080-kube-api-access-msdhj\") pod \"catalog-operator-68c6474976-gt2sl\" (UID: \"3b1fec22-5e5d-46d8-82ae-03440a6df080\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.619786 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.631282 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.635437 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.641067 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l"] Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.650798 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.663673 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22d364b4-96d8-46fa-b66f-57eabab1d389-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qqj26\" (UID: \"22d364b4-96d8-46fa-b66f-57eabab1d389\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:01 crc kubenswrapper[4759]: W1125 19:36:01.673963 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54c4b203_e670_4bca_9547_1cdbcaa58632.slice/crio-f1d78f121576754d5a5f4d8272eb55d46fdcab612cfe2789f699b9a229f64173 WatchSource:0}: Error finding container f1d78f121576754d5a5f4d8272eb55d46fdcab612cfe2789f699b9a229f64173: Status 404 returned error can't find the container with id f1d78f121576754d5a5f4d8272eb55d46fdcab612cfe2789f699b9a229f64173 Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.683607 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbljz\" (UniqueName: \"kubernetes.io/projected/f398f4ae-52d7-404e-94e8-84352c38f62a-kube-api-access-cbljz\") pod \"service-ca-9c57cc56f-mq5jf\" (UID: \"f398f4ae-52d7-404e-94e8-84352c38f62a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.702190 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6fgw\" (UniqueName: \"kubernetes.io/projected/ef151ed2-118a-4cb8-b197-ae81db91ef3c-kube-api-access-d6fgw\") pod \"machine-config-operator-74547568cd-lkwxn\" (UID: \"ef151ed2-118a-4cb8-b197-ae81db91ef3c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.713673 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mqv76"] Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.717345 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.738108 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljjzr\" (UniqueName: \"kubernetes.io/projected/9e2deb6a-8999-4f7d-a50a-7a3725e6a873-kube-api-access-ljjzr\") pod \"cluster-image-registry-operator-dc59b4c8b-dwpwk\" (UID: \"9e2deb6a-8999-4f7d-a50a-7a3725e6a873\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.738124 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cht75\" (UniqueName: \"kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75\") pod \"marketplace-operator-79b997595-kwszg\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.749214 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.758315 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.760235 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qg7bm"] Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.766606 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.772695 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcqc6\" (UniqueName: \"kubernetes.io/projected/9e729de1-5c72-4804-8f28-67bb35e37b50-kube-api-access-lcqc6\") pod \"multus-admission-controller-857f4d67dd-tkwdq\" (UID: \"9e729de1-5c72-4804-8f28-67bb35e37b50\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.773340 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.784430 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5fmd\" (UniqueName: \"kubernetes.io/projected/2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5-kube-api-access-d5fmd\") pod \"router-default-5444994796-k7nmp\" (UID: \"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5\") " pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.787981 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.798499 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj982\" (UniqueName: \"kubernetes.io/projected/5006b06d-5866-4d7a-a6e6-e68d8ed9ef77-kube-api-access-wj982\") pod \"apiserver-76f77b778f-vzt92\" (UID: \"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77\") " pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.798711 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkz9g"] Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.800152 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.815540 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce66674d-c2cc-4b29-87fb-33474d19da87-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kxk99\" (UID: \"ce66674d-c2cc-4b29-87fb-33474d19da87\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:36:01 crc kubenswrapper[4759]: W1125 19:36:01.830985 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d156e7_cd25_4e40_a754_f9dc88fbf01f.slice/crio-1d7379fff6ee7c057aabcd3bdd312f2cb343b3e54275177d15645139bf83a66f WatchSource:0}: Error finding container 1d7379fff6ee7c057aabcd3bdd312f2cb343b3e54275177d15645139bf83a66f: Status 404 returned error can't find the container with id 1d7379fff6ee7c057aabcd3bdd312f2cb343b3e54275177d15645139bf83a66f Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.832905 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.846090 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.854095 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht4nl\" (UniqueName: \"kubernetes.io/projected/47052dc0-053f-422a-814b-09d6d56f0a6d-kube-api-access-ht4nl\") pod \"dns-operator-744455d44c-rfwts\" (UID: \"47052dc0-053f-422a-814b-09d6d56f0a6d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.858657 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.870370 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.882697 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn8c9\" (UniqueName: \"kubernetes.io/projected/64718f2c-6468-414f-a80f-a44ab3ec63a2-kube-api-access-xn8c9\") pod \"service-ca-operator-777779d784-dpc7z\" (UID: \"64718f2c-6468-414f-a80f-a44ab3ec63a2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.885109 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzcjk\" (UniqueName: \"kubernetes.io/projected/b9999873-da0d-4ed3-88ac-cdfd8865cdd2-kube-api-access-xzcjk\") pod \"dns-default-mqn2l\" (UID: \"b9999873-da0d-4ed3-88ac-cdfd8865cdd2\") " pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.907635 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xhjh\" (UniqueName: \"kubernetes.io/projected/da4f223e-5556-440d-93ad-eacfc180ae21-kube-api-access-8xhjh\") pod \"ingress-canary-7zhk5\" (UID: \"da4f223e-5556-440d-93ad-eacfc180ae21\") " pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.909486 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" event={"ID":"54c4b203-e670-4bca-9547-1cdbcaa58632","Type":"ContainerStarted","Data":"f1d78f121576754d5a5f4d8272eb55d46fdcab612cfe2789f699b9a229f64173"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.910563 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" event={"ID":"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6","Type":"ContainerStarted","Data":"9857afd8dd4a4d12e673700dbe5e86bc6213dc37822febd9322c022753cf6ad1"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.912659 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4szh4" event={"ID":"32444b14-e189-4a50-8e25-ed6323de94c7","Type":"ContainerStarted","Data":"956d0ed87eb76f8d90d7d41497214c570d152524bcf5a2cf377c442e91f12598"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.912695 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4szh4" event={"ID":"32444b14-e189-4a50-8e25-ed6323de94c7","Type":"ContainerStarted","Data":"e877f03a419e88510b4e40c0ab0346cd732fb43ab66cd0c05167c3eac3e68983"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.913744 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" event={"ID":"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16","Type":"ContainerStarted","Data":"cf1d569709e5b60e590a9150bcdabd378de3bed61e8debf376c933346edb0928"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.915550 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" event={"ID":"65d156e7-cd25-4e40-a754-f9dc88fbf01f","Type":"ContainerStarted","Data":"1d7379fff6ee7c057aabcd3bdd312f2cb343b3e54275177d15645139bf83a66f"} Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.926399 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.927037 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7c6d\" (UniqueName: \"kubernetes.io/projected/1e8f64c9-2b3b-4e0f-93ea-78f589d591c5-kube-api-access-r7c6d\") pod \"machine-config-controller-84d6567774-6hvlv\" (UID: \"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.934872 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h26sc\" (UniqueName: \"kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc\") pod \"collect-profiles-29401650-lfdkx\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.939890 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.957151 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.960475 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hwk7\" (UniqueName: \"kubernetes.io/projected/6bcc761d-f5bd-476b-a8e5-25dcca68942e-kube-api-access-4hwk7\") pod \"csi-hostpathplugin-wrlnm\" (UID: \"6bcc761d-f5bd-476b-a8e5-25dcca68942e\") " pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:36:01 crc kubenswrapper[4759]: I1125 19:36:01.990416 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.004842 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.029930 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p85dv\" (UniqueName: \"kubernetes.io/projected/33cde591-0bb8-411a-bb3f-9fb30b730063-kube-api-access-p85dv\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030006 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030031 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-node-bootstrap-token\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030072 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030097 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33cde591-0bb8-411a-bb3f-9fb30b730063-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030122 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030143 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbnnd\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030166 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4h9s\" (UniqueName: \"kubernetes.io/projected/585a28b2-6137-42b0-8f6b-c8cf159b65db-kube-api-access-c4h9s\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030208 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3eee6397-9eb2-4816-bd39-71bab19aa521-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030269 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-certs\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030298 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3eee6397-9eb2-4816-bd39-71bab19aa521-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030350 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030386 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030473 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030521 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.030623 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eee6397-9eb2-4816-bd39-71bab19aa521-config\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.035278 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:02.535266534 +0000 UTC m=+153.121214235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.128800 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.131612 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.131737 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3eee6397-9eb2-4816-bd39-71bab19aa521-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.131816 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.131878 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.131972 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132254 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eee6397-9eb2-4816-bd39-71bab19aa521-config\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132411 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p85dv\" (UniqueName: \"kubernetes.io/projected/33cde591-0bb8-411a-bb3f-9fb30b730063-kube-api-access-p85dv\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132487 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132510 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-node-bootstrap-token\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132607 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132633 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33cde591-0bb8-411a-bb3f-9fb30b730063-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132704 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132726 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbnnd\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132750 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4h9s\" (UniqueName: \"kubernetes.io/projected/585a28b2-6137-42b0-8f6b-c8cf159b65db-kube-api-access-c4h9s\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132876 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3eee6397-9eb2-4816-bd39-71bab19aa521-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.132894 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-certs\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.134289 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:02.634244159 +0000 UTC m=+153.220191960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.134478 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.135031 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.136426 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eee6397-9eb2-4816-bd39-71bab19aa521-config\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.136643 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.140893 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.149002 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7zhk5" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.152999 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.153026 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-node-bootstrap-token\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.153676 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3eee6397-9eb2-4816-bd39-71bab19aa521-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.155269 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.155272 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/585a28b2-6137-42b0-8f6b-c8cf159b65db-certs\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.157844 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33cde591-0bb8-411a-bb3f-9fb30b730063-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.168749 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.168783 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-x8rv5"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.168792 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.177028 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.179992 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.180513 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dnkhr"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.213332 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.225269 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbnnd\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.234244 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4h9s\" (UniqueName: \"kubernetes.io/projected/585a28b2-6137-42b0-8f6b-c8cf159b65db-kube-api-access-c4h9s\") pod \"machine-config-server-wcq7p\" (UID: \"585a28b2-6137-42b0-8f6b-c8cf159b65db\") " pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.234600 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.234880 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:02.734868549 +0000 UTC m=+153.320816250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.248945 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3eee6397-9eb2-4816-bd39-71bab19aa521-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wf96k\" (UID: \"3eee6397-9eb2-4816-bd39-71bab19aa521\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.271510 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p85dv\" (UniqueName: \"kubernetes.io/projected/33cde591-0bb8-411a-bb3f-9fb30b730063-kube-api-access-p85dv\") pod \"package-server-manager-789f6589d5-7skhl\" (UID: \"33cde591-0bb8-411a-bb3f-9fb30b730063\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.315863 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.318637 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pvn5h"] Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.339181 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.340001 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:02.839980282 +0000 UTC m=+153.425927983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.341038 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz"] Nov 25 19:36:02 crc kubenswrapper[4759]: W1125 19:36:02.438071 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod720df419_4c33_4e7f_9bfd_147191aa10e6.slice/crio-1018bbb96b1ac3cf63f7d40a213539af226bcd375e0f41db8765e6589fe4dcb1 WatchSource:0}: Error finding container 1018bbb96b1ac3cf63f7d40a213539af226bcd375e0f41db8765e6589fe4dcb1: Status 404 returned error can't find the container with id 1018bbb96b1ac3cf63f7d40a213539af226bcd375e0f41db8765e6589fe4dcb1 Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.442150 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.442389 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:02.94237846 +0000 UTC m=+153.528326161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.497065 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wcq7p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.514685 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.531774 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.543568 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.543666 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.043646247 +0000 UTC m=+153.629593948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.543918 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.544295 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.044280225 +0000 UTC m=+153.630227926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.645081 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.645874 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.145854231 +0000 UTC m=+153.731801932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.747313 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.747635 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.247621952 +0000 UTC m=+153.833569653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.848266 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.848547 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.3485263 +0000 UTC m=+153.934474001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.848845 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.849296 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.34927808 +0000 UTC m=+153.935225781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.924554 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" event={"ID":"65d156e7-cd25-4e40-a754-f9dc88fbf01f","Type":"ContainerStarted","Data":"3c14d14a7f5377b11f16ea20e23a480e03aaa72b58b1c2972e0cfda36f058e34"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.934348 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" event={"ID":"540c9b5f-650e-4708-a665-ed3a7ea52b70","Type":"ContainerStarted","Data":"c85a932136713b882fdfcc67db08d255aa27123de5d788f6e353e87cd673017d"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.934388 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" event={"ID":"540c9b5f-650e-4708-a665-ed3a7ea52b70","Type":"ContainerStarted","Data":"6aef254d334552b560ea083d22e1f82fd3c7f0dc40248b06e76a0a0207db1e2e"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.935641 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.948371 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" event={"ID":"a53e86aa-8a32-40af-a371-0055d170c5ce","Type":"ContainerStarted","Data":"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.948424 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" event={"ID":"a53e86aa-8a32-40af-a371-0055d170c5ce","Type":"ContainerStarted","Data":"3f6a9ba8db68876f4db4feda608c526de683aa2e356535c6c54ccdddee549248"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.948855 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.949550 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:02 crc kubenswrapper[4759]: E1125 19:36:02.949893 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.449855119 +0000 UTC m=+154.035802820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.951483 4759 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ddq5p container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.951552 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.953646 4759 generic.go:334] "Generic (PLEG): container finished" podID="720df419-4c33-4e7f-9bfd-147191aa10e6" containerID="97bb757fa4b49e3554b7613ac64088f763b82e92f2a878d4e38c61bbfca6850b" exitCode=0 Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.953721 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" event={"ID":"720df419-4c33-4e7f-9bfd-147191aa10e6","Type":"ContainerDied","Data":"97bb757fa4b49e3554b7613ac64088f763b82e92f2a878d4e38c61bbfca6850b"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.953751 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" event={"ID":"720df419-4c33-4e7f-9bfd-147191aa10e6","Type":"ContainerStarted","Data":"1018bbb96b1ac3cf63f7d40a213539af226bcd375e0f41db8765e6589fe4dcb1"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.962972 4759 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fcwxs container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.963146 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" podUID="540c9b5f-650e-4708-a665-ed3a7ea52b70" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.978804 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wcq7p" event={"ID":"585a28b2-6137-42b0-8f6b-c8cf159b65db","Type":"ContainerStarted","Data":"2bd777985c3af93dcd1c297aea0ed7f32aae022ade5fb58409b4e01882f115f4"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.978849 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wcq7p" event={"ID":"585a28b2-6137-42b0-8f6b-c8cf159b65db","Type":"ContainerStarted","Data":"8729439b076aad1eeb08b4cbd18c3cbaa63da90d6893ec00554a55f7a7f5c4c8"} Nov 25 19:36:02 crc kubenswrapper[4759]: I1125 19:36:02.985541 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" event={"ID":"54c4b203-e670-4bca-9547-1cdbcaa58632","Type":"ContainerStarted","Data":"dbb62beb1f865f2dd95ecd26e6e25df8611098632050fb611360d2b666837290"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.008072 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" event={"ID":"d74b7f2c-ab7c-4618-a700-5b43b5e65cb6","Type":"ContainerStarted","Data":"a630e2aff4cc23379c799b1fb63755a4deb03db1d428a906d0f020283fc27ab9"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.009069 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.012831 4759 patch_prober.go:28] interesting pod/console-operator-58897d9998-qg7bm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.012944 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" podUID="d74b7f2c-ab7c-4618-a700-5b43b5e65cb6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.019858 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" event={"ID":"1bc8d591-5c1b-4703-9967-99f84f455cd4","Type":"ContainerStarted","Data":"9fd1ee4791e346d9f8a69d2f9cd93adc069f6dc98cd2ebb2a347ef71c8958480"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.019902 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" event={"ID":"1bc8d591-5c1b-4703-9967-99f84f455cd4","Type":"ContainerStarted","Data":"a95d179cf873520a5425bbb40acb60d754b6fdb1073108fd87b4e7081b08188a"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.028068 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" event={"ID":"e5229cef-43ed-49af-a71f-e389bc154e8d","Type":"ContainerStarted","Data":"8f5e1db5368e33bbe854c1b541debe7b95d6efbbd888ec8ccdb1dd0337e3711a"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.034345 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k7nmp" event={"ID":"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5","Type":"ContainerStarted","Data":"c3257bdc75e6971efe349d78fc004bd78a36b7f00e5116e3faf7ba3228a2442f"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.034388 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k7nmp" event={"ID":"2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5","Type":"ContainerStarted","Data":"7a12a82ef010812223c2b3c6e0477c903ee245fd28968cd21bd0f83465e55a55"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.038010 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" event={"ID":"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0","Type":"ContainerStarted","Data":"b784f094ad6369d8ce4bdee4e3b06689d0ec68c2e65c2934df19b44b7f10fd5d"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.038056 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" event={"ID":"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0","Type":"ContainerStarted","Data":"7bf6a21cc21df0bc844c94891e8c7996a1a3eba6030fba3f3395d3127f25f7db"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.039427 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.040159 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-x8rv5" event={"ID":"4e9ac029-8117-49ff-be41-19f18e9f621a","Type":"ContainerStarted","Data":"f5f4c285888e6ca3b433391cc1a3f0dc058e54cdaf2a6bd11904a640cf71d44a"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.040195 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-x8rv5" event={"ID":"4e9ac029-8117-49ff-be41-19f18e9f621a","Type":"ContainerStarted","Data":"634af34635ea1f97d01db9cff431c705d27bb64876a08e6adb0af97088d2da79"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.040481 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.041824 4759 patch_prober.go:28] interesting pod/downloads-7954f5f757-x8rv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.041861 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-x8rv5" podUID="4e9ac029-8117-49ff-be41-19f18e9f621a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.041885 4759 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pvn5h container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.041927 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.049882 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" event={"ID":"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16","Type":"ContainerStarted","Data":"17cef7a2318408a85ec5ae84dfdcc25c050439365bf4e881a98ed0de5fa54272"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.049919 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" event={"ID":"0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16","Type":"ContainerStarted","Data":"e7529ed48c94fd7b9e324b8cc923db05a36748506f631f85449087cc2fad8e56"} Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.055270 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.056205 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.556187954 +0000 UTC m=+154.142135655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.153631 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.159434 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.171289 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.671251497 +0000 UTC m=+154.257199198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.192897 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tkwdq"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.198176 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.236795 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p6xcd" podStartSLOduration=133.236775882 podStartE2EDuration="2m13.236775882s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.233256076 +0000 UTC m=+153.819203777" watchObservedRunningTime="2025-11-25 19:36:03.236775882 +0000 UTC m=+153.822723583" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.264339 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.273163 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.773149503 +0000 UTC m=+154.359097204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.311124 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" podStartSLOduration=132.311106976 podStartE2EDuration="2m12.311106976s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.308897155 +0000 UTC m=+153.894844856" watchObservedRunningTime="2025-11-25 19:36:03.311106976 +0000 UTC m=+153.897054697" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.357080 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l98cg" podStartSLOduration=132.357060948 podStartE2EDuration="2m12.357060948s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.352989456 +0000 UTC m=+153.938937157" watchObservedRunningTime="2025-11-25 19:36:03.357060948 +0000 UTC m=+153.943008659" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.365718 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.366096 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.866078193 +0000 UTC m=+154.452025894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.465476 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.467386 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.467691 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:03.96767979 +0000 UTC m=+154.553627491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.496138 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.508579 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.517240 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.518378 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn"] Nov 25 19:36:03 crc kubenswrapper[4759]: W1125 19:36:03.519791 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac427820_58ca_4e4f_b8c6_058084c79617.slice/crio-271e0baa3eea51ecb3de3590c4a7d2072d0159b805e814457fd7de870df4a7ba WatchSource:0}: Error finding container 271e0baa3eea51ecb3de3590c4a7d2072d0159b805e814457fd7de870df4a7ba: Status 404 returned error can't find the container with id 271e0baa3eea51ecb3de3590c4a7d2072d0159b805e814457fd7de870df4a7ba Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.536564 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7zhk5"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.568061 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.568341 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.06832605 +0000 UTC m=+154.654273751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.584188 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh"] Nov 25 19:36:03 crc kubenswrapper[4759]: W1125 19:36:03.596946 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef151ed2_118a_4cb8_b197_ae81db91ef3c.slice/crio-798ba9b4d4bd741307e14f503a546bd06eeb7c22770587f2688c8e1dcd685f62 WatchSource:0}: Error finding container 798ba9b4d4bd741307e14f503a546bd06eeb7c22770587f2688c8e1dcd685f62: Status 404 returned error can't find the container with id 798ba9b4d4bd741307e14f503a546bd06eeb7c22770587f2688c8e1dcd685f62 Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.610994 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.614554 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.623471 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mqn2l"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.633837 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mq5jf"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.649429 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.652058 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.654590 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.657184 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.658652 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfwts"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.663209 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl"] Nov 25 19:36:03 crc kubenswrapper[4759]: W1125 19:36:03.678428 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22d364b4_96d8_46fa_b66f_57eabab1d389.slice/crio-5280d72b0014442aaa61bd176346ec7e4c821bc6873e2429fd95fd84d289d80e WatchSource:0}: Error finding container 5280d72b0014442aaa61bd176346ec7e4c821bc6873e2429fd95fd84d289d80e: Status 404 returned error can't find the container with id 5280d72b0014442aaa61bd176346ec7e4c821bc6873e2429fd95fd84d289d80e Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.687949 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.688390 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.18837568 +0000 UTC m=+154.774323381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.689587 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.716786 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vzt92"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.741769 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wrlnm"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.749330 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99"] Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.750181 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4szh4" podStartSLOduration=133.750155802 podStartE2EDuration="2m13.750155802s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.635931841 +0000 UTC m=+154.221879542" watchObservedRunningTime="2025-11-25 19:36:03.750155802 +0000 UTC m=+154.336103503" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.790358 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.790723 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.290676975 +0000 UTC m=+154.876624676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.790875 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.791319 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.291303763 +0000 UTC m=+154.877251464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.844400 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mqv76" podStartSLOduration=132.844379398 podStartE2EDuration="2m12.844379398s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.843279148 +0000 UTC m=+154.429226849" watchObservedRunningTime="2025-11-25 19:36:03.844379398 +0000 UTC m=+154.430327099" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.847010 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.879983 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:03 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:03 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:03 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.880029 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.899643 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:03 crc kubenswrapper[4759]: E1125 19:36:03.900120 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.400103545 +0000 UTC m=+154.986051246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:03 crc kubenswrapper[4759]: I1125 19:36:03.914308 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wcq7p" podStartSLOduration=5.914288812 podStartE2EDuration="5.914288812s" podCreationTimestamp="2025-11-25 19:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:03.914015453 +0000 UTC m=+154.499963154" watchObservedRunningTime="2025-11-25 19:36:03.914288812 +0000 UTC m=+154.500236513" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.002112 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.002499 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.502486073 +0000 UTC m=+155.088433774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.102844 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.103144 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.603104953 +0000 UTC m=+155.189052654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.103330 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.104042 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.604034718 +0000 UTC m=+155.189982419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.104181 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" event={"ID":"6bcc761d-f5bd-476b-a8e5-25dcca68942e","Type":"ContainerStarted","Data":"ade6dd9225fb8aa8c8418d8942d69f83edcd87005c2ba029db38ef85327ceb42"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.120646 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" event={"ID":"22d364b4-96d8-46fa-b66f-57eabab1d389","Type":"ContainerStarted","Data":"5280d72b0014442aaa61bd176346ec7e4c821bc6873e2429fd95fd84d289d80e"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.134759 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" event={"ID":"3eee6397-9eb2-4816-bd39-71bab19aa521","Type":"ContainerStarted","Data":"ba829b39a3c9954ac7abd9f6efa2f03093ff7babc90f6c93c989ec5fbdf27536"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.148667 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" event={"ID":"48a44627-b436-4b88-bab8-0190ab7493e6","Type":"ContainerStarted","Data":"d9836dc2c6a9d2420ad442caa4ee094e0a248b9e11eab35dc5929b43c4e820db"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.149107 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" event={"ID":"48a44627-b436-4b88-bab8-0190ab7493e6","Type":"ContainerStarted","Data":"47e3084f2addf5dd3eb491b8dd4f9dbad35f624acd115c092db678f77add8143"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.153607 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" event={"ID":"ac427820-58ca-4e4f-b8c6-058084c79617","Type":"ContainerStarted","Data":"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.153645 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.153657 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" event={"ID":"ac427820-58ca-4e4f-b8c6-058084c79617","Type":"ContainerStarted","Data":"271e0baa3eea51ecb3de3590c4a7d2072d0159b805e814457fd7de870df4a7ba"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.159996 4759 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kwszg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.160055 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.160283 4759 generic.go:334] "Generic (PLEG): container finished" podID="124dfa3b-114d-4d88-961e-8b99f5530562" containerID="3536725ec7ad828a30f5411be41f6ba3c469b72ba1d08fd735ca3f813a407374" exitCode=0 Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.160360 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" event={"ID":"124dfa3b-114d-4d88-961e-8b99f5530562","Type":"ContainerDied","Data":"3536725ec7ad828a30f5411be41f6ba3c469b72ba1d08fd735ca3f813a407374"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.160391 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" event={"ID":"124dfa3b-114d-4d88-961e-8b99f5530562","Type":"ContainerStarted","Data":"5c686627c656623182856157bd0010da64fad18d9d75276ff6d9ff856d65b513"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.164373 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7zhk5" event={"ID":"da4f223e-5556-440d-93ad-eacfc180ae21","Type":"ContainerStarted","Data":"59b9017c9b020ee918acd53c3ebb5563d1e6f42efea84d2ab3dbfb5f0fede304"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.166577 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" event={"ID":"9e2deb6a-8999-4f7d-a50a-7a3725e6a873","Type":"ContainerStarted","Data":"240cf5118c201cd9341ae332c8ff54163a22671a606d6ec39cf4dc4876fe8963"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.168380 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" event={"ID":"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb","Type":"ContainerStarted","Data":"dfb77fa48ccc7120436748d91e8480707c72f02793486085ef34d5af3979508c"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.168408 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" event={"ID":"32fdca48-e90c-4f5e-8524-3abaeb2f1bfb","Type":"ContainerStarted","Data":"6d3d13c90818169f4e3c4b9cc8918dbe58aaf3beea8ffecbf29cb4e9a37cb0fc"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.170790 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" event={"ID":"35063c0f-1f00-4f84-9d90-986955cf006e","Type":"ContainerStarted","Data":"a2b989a3165a9dc480c50723ffdf31d5e829ec53743ec5276aa03d9a9d6b5817"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.172480 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" event={"ID":"e5229cef-43ed-49af-a71f-e389bc154e8d","Type":"ContainerStarted","Data":"443eb668e727da7487f55b06db841f121709b1838d1a2d9425a211866b56c46c"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.172507 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" event={"ID":"e5229cef-43ed-49af-a71f-e389bc154e8d","Type":"ContainerStarted","Data":"d9a6858129e3d808cab3bf40a984573cbe50b2f2e553d478e1dd7519f253e4a2"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.174406 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" event={"ID":"9e729de1-5c72-4804-8f28-67bb35e37b50","Type":"ContainerStarted","Data":"1e66d489969d93a2cb2aa76252077fd07985821b1cab4f08be1a9ac38312c12c"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.174476 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" event={"ID":"9e729de1-5c72-4804-8f28-67bb35e37b50","Type":"ContainerStarted","Data":"a21f7cd94d4d94b3400367d980515f3ab8228b709d4460b35b08f5edcb726ac4"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.204136 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.204308 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.704285358 +0000 UTC m=+155.290233059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.204517 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.206383 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.706373165 +0000 UTC m=+155.292320866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.215871 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" event={"ID":"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5","Type":"ContainerStarted","Data":"44b1d0193fe14c254a3a6b6ee48d473341ce37c7d0c2b9845e6068fd6d07f453"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.274273 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" event={"ID":"a0289180-1fc0-4d66-86ce-4e09fee7e0a9","Type":"ContainerStarted","Data":"18bfcabb32ad86abd878e7e940d829c71104702ca1658946a910f46d17f6b4d4"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.274358 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" event={"ID":"a0289180-1fc0-4d66-86ce-4e09fee7e0a9","Type":"ContainerStarted","Data":"71ddbe02c49e9c3fa93e4da8b9f3522691d054b36a5c977b523344848dc402a3"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.276684 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" event={"ID":"3b1fec22-5e5d-46d8-82ae-03440a6df080","Type":"ContainerStarted","Data":"d56031dd386137494abcf1c7c07314afbb6ed9d2345e30f5e2a0e4a283f2129f"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.282899 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" event={"ID":"64718f2c-6468-414f-a80f-a44ab3ec63a2","Type":"ContainerStarted","Data":"c7ad63ee1f18963ff3aee5191855a3f4e62ea5034ba80864e061a50dce84abe3"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.307745 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" event={"ID":"ce66674d-c2cc-4b29-87fb-33474d19da87","Type":"ContainerStarted","Data":"6f3c2d62a7a2c809215a62c360db5d58160190922049fc2f9fb3fe014a8ac288"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.308832 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.309704 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.809681739 +0000 UTC m=+155.395629440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.311104 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.311927 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.811911428 +0000 UTC m=+155.397859129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.314781 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" event={"ID":"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77","Type":"ContainerStarted","Data":"48ce1d11e30e249ce09219445e04d8a041cdae1bd58a1dd37903bde5c8fc1f4f"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.328924 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" event={"ID":"33cde591-0bb8-411a-bb3f-9fb30b730063","Type":"ContainerStarted","Data":"43f0446b6d91a9abfced49ca09dc135b2de5ef9973d0c22c33f2b03908f29106"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.332053 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-k7nmp" podStartSLOduration=133.332042828 podStartE2EDuration="2m13.332042828s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.331021989 +0000 UTC m=+154.916969690" watchObservedRunningTime="2025-11-25 19:36:04.332042828 +0000 UTC m=+154.917990529" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.353318 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-x8rv5" podStartSLOduration=134.353286955 podStartE2EDuration="2m14.353286955s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.351421935 +0000 UTC m=+154.937369636" watchObservedRunningTime="2025-11-25 19:36:04.353286955 +0000 UTC m=+154.939234646" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.355295 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" event={"ID":"47052dc0-053f-422a-814b-09d6d56f0a6d","Type":"ContainerStarted","Data":"c6eec0bed24bd1c806090bda0d8b15935abe834086f5a4eed832a54e747b46e9"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.364386 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mqn2l" event={"ID":"b9999873-da0d-4ed3-88ac-cdfd8865cdd2","Type":"ContainerStarted","Data":"652165bf8fdea35e45c547cba0bdec01595ee5e46460cefc685a38eccf27852f"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.417428 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.417960 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:04.917927636 +0000 UTC m=+155.503875337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.442347 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rnw7l" podStartSLOduration=134.44233145 podStartE2EDuration="2m14.44233145s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.397168941 +0000 UTC m=+154.983116642" watchObservedRunningTime="2025-11-25 19:36:04.44233145 +0000 UTC m=+155.028279151" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.469797 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" event={"ID":"ef151ed2-118a-4cb8-b197-ae81db91ef3c","Type":"ContainerStarted","Data":"ff864be8ea32f5a91322c3525e3427e2410d4619f5e5404867dd6043bf89a872"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.469844 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" event={"ID":"ef151ed2-118a-4cb8-b197-ae81db91ef3c","Type":"ContainerStarted","Data":"798ba9b4d4bd741307e14f503a546bd06eeb7c22770587f2688c8e1dcd685f62"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.487879 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" event={"ID":"720df419-4c33-4e7f-9bfd-147191aa10e6","Type":"ContainerStarted","Data":"040793a4552dbd3e64dc202be25dc4087a6d9c172eed4b0b11681cd4833c45d0"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.514540 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" event={"ID":"0c794333-1489-4d9a-b395-3648650ede06","Type":"ContainerStarted","Data":"1782dd2c8b1a73bb7371fc2a2ed40df5b7959a86117f9f990f51e98f6e7240a0"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.514579 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" event={"ID":"0c794333-1489-4d9a-b395-3648650ede06","Type":"ContainerStarted","Data":"92f672107eb01d46c08eb3a4d2fe5b349cdb32d7da3f8153fe15c2922d2c19f7"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.514786 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.523772 4759 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-w5jkg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.523826 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" podUID="0c794333-1489-4d9a-b395-3648650ede06" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.524881 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" podStartSLOduration=133.524864049 podStartE2EDuration="2m13.524864049s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.522553576 +0000 UTC m=+155.108501287" watchObservedRunningTime="2025-11-25 19:36:04.524864049 +0000 UTC m=+155.110811750" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.532351 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.533674 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.033658748 +0000 UTC m=+155.619606449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.555870 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" event={"ID":"f398f4ae-52d7-404e-94e8-84352c38f62a","Type":"ContainerStarted","Data":"92798ff8478e74819d6775e3cd357243464e496b03d022a6d4d9ac6347d037eb"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.556511 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dnkhr" podStartSLOduration=133.55648994 podStartE2EDuration="2m13.55648994s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.552205892 +0000 UTC m=+155.138153593" watchObservedRunningTime="2025-11-25 19:36:04.55648994 +0000 UTC m=+155.142437641" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.606004 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" podStartSLOduration=134.605985487 podStartE2EDuration="2m14.605985487s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.604500747 +0000 UTC m=+155.190448458" watchObservedRunningTime="2025-11-25 19:36:04.605985487 +0000 UTC m=+155.191933188" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.633160 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.633393 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkz9g" podStartSLOduration=134.633372963 podStartE2EDuration="2m14.633372963s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.631202633 +0000 UTC m=+155.217150324" watchObservedRunningTime="2025-11-25 19:36:04.633372963 +0000 UTC m=+155.219320664" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.633519 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.133504036 +0000 UTC m=+155.719451747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.634961 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.636026 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.136012865 +0000 UTC m=+155.721960566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.696408 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" event={"ID":"a88a31d1-eda7-401b-b5e0-f14376043180","Type":"ContainerStarted","Data":"eae71f64b4882fec7b0d7ca274e139f2f7f5d24863c07e4e6bda611e13864bcf"} Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.697247 4759 patch_prober.go:28] interesting pod/downloads-7954f5f757-x8rv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.697269 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-x8rv5" podUID="4e9ac029-8117-49ff-be41-19f18e9f621a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.704570 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.705315 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fcwxs" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.707759 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.733139 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.736419 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.736605 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.236583673 +0000 UTC m=+155.822531374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.736702 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.737922 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.23790729 +0000 UTC m=+155.823854991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.854959 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.856554 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.35653215 +0000 UTC m=+155.942479851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.863506 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:04 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:04 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:04 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.863581 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.934761 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qg7bm" podStartSLOduration=134.9347379 podStartE2EDuration="2m14.9347379s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.934151634 +0000 UTC m=+155.520099345" watchObservedRunningTime="2025-11-25 19:36:04.9347379 +0000 UTC m=+155.520685601" Nov 25 19:36:04 crc kubenswrapper[4759]: I1125 19:36:04.964112 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:04 crc kubenswrapper[4759]: E1125 19:36:04.964476 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.464464528 +0000 UTC m=+156.050412229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.012944 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" podStartSLOduration=134.012916898 podStartE2EDuration="2m14.012916898s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:04.967292015 +0000 UTC m=+155.553239726" watchObservedRunningTime="2025-11-25 19:36:05.012916898 +0000 UTC m=+155.598864599" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.064771 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.064994 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.564963815 +0000 UTC m=+156.150911516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.065139 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.065476 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.565463909 +0000 UTC m=+156.151411610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.165208 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8bf6n" podStartSLOduration=134.165185894 podStartE2EDuration="2m14.165185894s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.111889463 +0000 UTC m=+155.697837164" watchObservedRunningTime="2025-11-25 19:36:05.165185894 +0000 UTC m=+155.751133595" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.173241 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.173621 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.673605124 +0000 UTC m=+156.259552825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.203188 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b9tgz" podStartSLOduration=135.203167909 podStartE2EDuration="2m15.203167909s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.201180985 +0000 UTC m=+155.787128686" watchObservedRunningTime="2025-11-25 19:36:05.203167909 +0000 UTC m=+155.789115610" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.278199 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" podStartSLOduration=134.278182301 podStartE2EDuration="2m14.278182301s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.277255036 +0000 UTC m=+155.863202737" watchObservedRunningTime="2025-11-25 19:36:05.278182301 +0000 UTC m=+155.864130002" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.287747 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.288346 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.788329908 +0000 UTC m=+156.374277609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.361163 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" podStartSLOduration=134.36114175 podStartE2EDuration="2m14.36114175s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.360809281 +0000 UTC m=+155.946756982" watchObservedRunningTime="2025-11-25 19:36:05.36114175 +0000 UTC m=+155.947089441" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.395107 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" podStartSLOduration=135.395092125 podStartE2EDuration="2m15.395092125s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.393471091 +0000 UTC m=+155.979418792" watchObservedRunningTime="2025-11-25 19:36:05.395092125 +0000 UTC m=+155.981039826" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.396687 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.396881 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.896867434 +0000 UTC m=+156.482815135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.396965 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.397307 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.897297995 +0000 UTC m=+156.483245696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.484087 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mlsn" podStartSLOduration=134.484054458 podStartE2EDuration="2m14.484054458s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.456837596 +0000 UTC m=+156.042785297" watchObservedRunningTime="2025-11-25 19:36:05.484054458 +0000 UTC m=+156.070002159" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.497696 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.497965 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:05.997950446 +0000 UTC m=+156.583898147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.599102 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.599428 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.099415399 +0000 UTC m=+156.685363100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.641423 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" podStartSLOduration=134.641407282 podStartE2EDuration="2m14.641407282s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.637371213 +0000 UTC m=+156.223318914" watchObservedRunningTime="2025-11-25 19:36:05.641407282 +0000 UTC m=+156.227354983" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.700967 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.701202 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.201167399 +0000 UTC m=+156.787115100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.701431 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.701826 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.201807347 +0000 UTC m=+156.787755048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.718192 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" event={"ID":"a88a31d1-eda7-401b-b5e0-f14376043180","Type":"ContainerStarted","Data":"c76325b9a6d88f12f2908027912dea98567daf36be250619c9a9a769ebafb137"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.718242 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" event={"ID":"a88a31d1-eda7-401b-b5e0-f14376043180","Type":"ContainerStarted","Data":"8b39605af51d7ca3ba50bb8b7808b07f79b4881f6b670ca60c4d4dc59760c916"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.732050 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" event={"ID":"9e2deb6a-8999-4f7d-a50a-7a3725e6a873","Type":"ContainerStarted","Data":"a41c08db991bf87585745f8d088165230dbb885985d6d16b6c5d66807e198216"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.738652 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" event={"ID":"33cde591-0bb8-411a-bb3f-9fb30b730063","Type":"ContainerStarted","Data":"049c9896561e29843fad7ddbfe3450018cf35d07e59270b79d43a44255ac7d70"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.738696 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" event={"ID":"33cde591-0bb8-411a-bb3f-9fb30b730063","Type":"ContainerStarted","Data":"3d51bc5922eac076cffe89c7b29fc643b091bfc655ddb8afe2ca788172b65def"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.739260 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.750221 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" event={"ID":"9e729de1-5c72-4804-8f28-67bb35e37b50","Type":"ContainerStarted","Data":"1c1d723276608613ba1083af6f3a6f851bbe980eaa48280527e8f3c8b526541e"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.758314 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" event={"ID":"47052dc0-053f-422a-814b-09d6d56f0a6d","Type":"ContainerStarted","Data":"0fd29d48ebb28af449ac6548e8a4cc1f4cc34cfebc73a9d592366e6a98cec841"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.758357 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" event={"ID":"47052dc0-053f-422a-814b-09d6d56f0a6d","Type":"ContainerStarted","Data":"05bc8ef9f52a779ab34136595f34b786f5b2bf1b759a5cceea0893ede4bf315d"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.761315 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" event={"ID":"64718f2c-6468-414f-a80f-a44ab3ec63a2","Type":"ContainerStarted","Data":"49db484ef053690ff34224d18e21ae79bb7a8b07f52b2c5c4035c4b03a7741a7"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.764509 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" event={"ID":"ef151ed2-118a-4cb8-b197-ae81db91ef3c","Type":"ContainerStarted","Data":"ea57282f903e185e1b93f2a7505ba661a846bb7dd7438197127475dfda508dae"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.771147 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" event={"ID":"35063c0f-1f00-4f84-9d90-986955cf006e","Type":"ContainerStarted","Data":"663099e800435f39888441b1f344de0d46612881e05062b87abc9f3c721b82de"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.774602 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" event={"ID":"f398f4ae-52d7-404e-94e8-84352c38f62a","Type":"ContainerStarted","Data":"f8e9a5c28913699825e592dcbfa11d2db5d49e56ba362fcf9b4d9ce773857530"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.779091 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" event={"ID":"3b1fec22-5e5d-46d8-82ae-03440a6df080","Type":"ContainerStarted","Data":"b7441158b44461e647aae0509c101ea70a6ac6230bece6ef9924d2fe23fe94ab"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.779885 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.782714 4759 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gt2sl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.782778 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" podUID="3b1fec22-5e5d-46d8-82ae-03440a6df080" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.788537 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dwpwk" podStartSLOduration=134.788516169 podStartE2EDuration="2m14.788516169s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.786563075 +0000 UTC m=+156.372510766" watchObservedRunningTime="2025-11-25 19:36:05.788516169 +0000 UTC m=+156.374463870" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.789072 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pb8fh" podStartSLOduration=134.789066504 podStartE2EDuration="2m14.789066504s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.75739801 +0000 UTC m=+156.343345711" watchObservedRunningTime="2025-11-25 19:36:05.789066504 +0000 UTC m=+156.375014205" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.789223 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7zhk5" event={"ID":"da4f223e-5556-440d-93ad-eacfc180ae21","Type":"ContainerStarted","Data":"346d65507aae1e5793b55e3c0324fca0aaa1bd99abf7ae5179d1372172c889f6"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.795266 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" event={"ID":"22d364b4-96d8-46fa-b66f-57eabab1d389","Type":"ContainerStarted","Data":"029924e5e01fa0b70d8bae7a9d9270ef41002b929478bcfb61b1d74e118884dd"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.795321 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" event={"ID":"22d364b4-96d8-46fa-b66f-57eabab1d389","Type":"ContainerStarted","Data":"169f6957b55461946db050aeec1c5e81078d9f06a89bca6815d76bf59b9c8d1f"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.805555 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" event={"ID":"3eee6397-9eb2-4816-bd39-71bab19aa521","Type":"ContainerStarted","Data":"4d9d58e018673846f4401d2d038a81d06513cc4bb2a02dea7a3d37410ab57402"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.805727 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.806389 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.306360334 +0000 UTC m=+156.892308035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.806591 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.807689 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" event={"ID":"ce66674d-c2cc-4b29-87fb-33474d19da87","Type":"ContainerStarted","Data":"192e50f1bcea55da4e50e568c6d21cf284612b4155c71f2fdcb443d220d1f8f4"} Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.807835 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.307818683 +0000 UTC m=+156.893766384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.812261 4759 generic.go:334] "Generic (PLEG): container finished" podID="5006b06d-5866-4d7a-a6e6-e68d8ed9ef77" containerID="cc6090dc16073d9d84cbbc818eff4a60f81eafcb43589e6ba552f40179459c39" exitCode=0 Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.812340 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" event={"ID":"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77","Type":"ContainerDied","Data":"cc6090dc16073d9d84cbbc818eff4a60f81eafcb43589e6ba552f40179459c39"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.815674 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mqn2l" event={"ID":"b9999873-da0d-4ed3-88ac-cdfd8865cdd2","Type":"ContainerStarted","Data":"b32a6a6a22e7509b0a68008a772f031cab6de8e5e97127aa94d971e591f4e2aa"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.815753 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mqn2l" event={"ID":"b9999873-da0d-4ed3-88ac-cdfd8865cdd2","Type":"ContainerStarted","Data":"f91992221a748328d0d6f41ea92a33fb82f510379dd650be20ca0446a28a7937"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.815889 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.829402 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" event={"ID":"124dfa3b-114d-4d88-961e-8b99f5530562","Type":"ContainerStarted","Data":"a0ed3a9849ef8af34c8cca625b7cf26510c7f72017b70c2cab238b64997dfbec"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.831129 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.850656 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" event={"ID":"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5","Type":"ContainerStarted","Data":"ca6d1d336cacf142cc1e98c864edcd0d29332de74e1a9058a43304ce58b5efba"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.850722 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" event={"ID":"1e8f64c9-2b3b-4e0f-93ea-78f589d591c5","Type":"ContainerStarted","Data":"0f7f15dc7cff854def6e3d25bd80327c4bea8bf85f7a9aa89aa30cb3eacf9949"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.856601 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:05 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:05 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:05 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.856845 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.878366 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tkwdq" podStartSLOduration=134.878347375 podStartE2EDuration="2m14.878347375s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.837289976 +0000 UTC m=+156.423237677" watchObservedRunningTime="2025-11-25 19:36:05.878347375 +0000 UTC m=+156.464295076" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.883131 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" event={"ID":"6bcc761d-f5bd-476b-a8e5-25dcca68942e","Type":"ContainerStarted","Data":"ee88e13a5f30a95b34557195fbbd2cb28627ffa78dbe14bd401d826943f6efdb"} Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.887293 4759 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kwszg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.887322 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.887366 4759 patch_prober.go:28] interesting pod/downloads-7954f5f757-x8rv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.887377 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-x8rv5" podUID="4e9ac029-8117-49ff-be41-19f18e9f621a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.914369 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:05 crc kubenswrapper[4759]: E1125 19:36:05.915612 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.415587329 +0000 UTC m=+157.001535030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.922799 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rfwts" podStartSLOduration=134.922783155 podStartE2EDuration="2m14.922783155s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.921861609 +0000 UTC m=+156.507809310" watchObservedRunningTime="2025-11-25 19:36:05.922783155 +0000 UTC m=+156.508730856" Nov 25 19:36:05 crc kubenswrapper[4759]: I1125 19:36:05.923867 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" podStartSLOduration=134.923860494 podStartE2EDuration="2m14.923860494s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.880075341 +0000 UTC m=+156.466023042" watchObservedRunningTime="2025-11-25 19:36:05.923860494 +0000 UTC m=+156.509808195" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.002130 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6hvlv" podStartSLOduration=135.002112875 podStartE2EDuration="2m15.002112875s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.001867518 +0000 UTC m=+156.587815219" watchObservedRunningTime="2025-11-25 19:36:06.002112875 +0000 UTC m=+156.588060576" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.002526 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lkwxn" podStartSLOduration=135.002519856 podStartE2EDuration="2m15.002519856s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:05.959570677 +0000 UTC m=+156.545518368" watchObservedRunningTime="2025-11-25 19:36:06.002519856 +0000 UTC m=+156.588467557" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.022898 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.026155 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.526140949 +0000 UTC m=+157.112088650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.042587 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dpc7z" podStartSLOduration=135.042571856 podStartE2EDuration="2m15.042571856s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.040720446 +0000 UTC m=+156.626668147" watchObservedRunningTime="2025-11-25 19:36:06.042571856 +0000 UTC m=+156.628519557" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.081089 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qqj26" podStartSLOduration=135.081070075 podStartE2EDuration="2m15.081070075s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.079527042 +0000 UTC m=+156.665474743" watchObservedRunningTime="2025-11-25 19:36:06.081070075 +0000 UTC m=+156.667017776" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.108127 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wf96k" podStartSLOduration=135.108111892 podStartE2EDuration="2m15.108111892s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.107758342 +0000 UTC m=+156.693706043" watchObservedRunningTime="2025-11-25 19:36:06.108111892 +0000 UTC m=+156.694059593" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.127812 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.128152 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.628137077 +0000 UTC m=+157.214084778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.143264 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-mq5jf" podStartSLOduration=135.143248798 podStartE2EDuration="2m15.143248798s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.140827002 +0000 UTC m=+156.726774703" watchObservedRunningTime="2025-11-25 19:36:06.143248798 +0000 UTC m=+156.729196499" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.170228 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mqn2l" podStartSLOduration=7.170213762 podStartE2EDuration="7.170213762s" podCreationTimestamp="2025-11-25 19:35:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.16938171 +0000 UTC m=+156.755329421" watchObservedRunningTime="2025-11-25 19:36:06.170213762 +0000 UTC m=+156.756161463" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.220009 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" podStartSLOduration=136.219992748 podStartE2EDuration="2m16.219992748s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.218999131 +0000 UTC m=+156.804946852" watchObservedRunningTime="2025-11-25 19:36:06.219992748 +0000 UTC m=+156.805940449" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.228846 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.230285 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.730274578 +0000 UTC m=+157.316222279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.281509 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9xhm8" podStartSLOduration=135.281492673 podStartE2EDuration="2m15.281492673s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.241091622 +0000 UTC m=+156.827039323" watchObservedRunningTime="2025-11-25 19:36:06.281492673 +0000 UTC m=+156.867440374" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.329717 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.329963 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.829947712 +0000 UTC m=+157.415895413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.351043 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kxk99" podStartSLOduration=135.351029726 podStartE2EDuration="2m15.351029726s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.347652164 +0000 UTC m=+156.933599865" watchObservedRunningTime="2025-11-25 19:36:06.351029726 +0000 UTC m=+156.936977427" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.351128 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" podStartSLOduration=135.351123798 podStartE2EDuration="2m15.351123798s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.315372595 +0000 UTC m=+156.901320296" watchObservedRunningTime="2025-11-25 19:36:06.351123798 +0000 UTC m=+156.937071499" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.432336 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.432696 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:06.93268523 +0000 UTC m=+157.518632921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.486722 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w5jkg" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.534716 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7zhk5" podStartSLOduration=7.534700728 podStartE2EDuration="7.534700728s" podCreationTimestamp="2025-11-25 19:35:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.394774717 +0000 UTC m=+156.980722418" watchObservedRunningTime="2025-11-25 19:36:06.534700728 +0000 UTC m=+157.120648429" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.538032 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.538401 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.038385898 +0000 UTC m=+157.624333599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.621684 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.621733 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.639825 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.640150 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.140139349 +0000 UTC m=+157.726087050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.652075 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.741024 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.741216 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.241189051 +0000 UTC m=+157.827136752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.741267 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.741585 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.241570561 +0000 UTC m=+157.827518262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.842430 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.842846 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.342830868 +0000 UTC m=+157.928778569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.848484 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:06 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:06 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:06 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.848537 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.888907 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" event={"ID":"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77","Type":"ContainerStarted","Data":"5cf9163d3a10e2598fa1ba6ca1259d89a0d165d118ab8b7d2144d69ba87e6541"} Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.888992 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" event={"ID":"5006b06d-5866-4d7a-a6e6-e68d8ed9ef77","Type":"ContainerStarted","Data":"5a30c7ea4523ae644f039c23513d315d55b4e5fc1243f4833e3ca475530cb5fc"} Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.898700 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r4wvg" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.926655 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.926713 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.929185 4759 patch_prober.go:28] interesting pod/apiserver-76f77b778f-vzt92 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.929238 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" podUID="5006b06d-5866-4d7a-a6e6-e68d8ed9ef77" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.932938 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gt2sl" Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.943405 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:06 crc kubenswrapper[4759]: E1125 19:36:06.943778 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.443762757 +0000 UTC m=+158.029710458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:06 crc kubenswrapper[4759]: I1125 19:36:06.954590 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" podStartSLOduration=136.954574931 podStartE2EDuration="2m16.954574931s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:06.941930497 +0000 UTC m=+157.527878198" watchObservedRunningTime="2025-11-25 19:36:06.954574931 +0000 UTC m=+157.540522622" Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.051500 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.051678 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.551653274 +0000 UTC m=+158.137600975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.052415 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.056398 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.556383203 +0000 UTC m=+158.142330904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.155958 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.156240 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.656226563 +0000 UTC m=+158.242174264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.262172 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.262486 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.762471816 +0000 UTC m=+158.348419517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.363581 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.363933 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.863919468 +0000 UTC m=+158.449867169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.465101 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.465408 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:07.965390361 +0000 UTC m=+158.551338062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.567101 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.567302 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.067276506 +0000 UTC m=+158.653224207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.567423 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.567703 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.067691357 +0000 UTC m=+158.653639058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.616516 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ngvqb" Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.668873 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.669032 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.168997486 +0000 UTC m=+158.754945187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.669183 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.669716 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.169700405 +0000 UTC m=+158.755648106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.770350 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.770738 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.270722946 +0000 UTC m=+158.856670647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.858936 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:07 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:07 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:07 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.859288 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.871316 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.871586 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.371576342 +0000 UTC m=+158.957524043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.889583 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.894963 4759 generic.go:334] "Generic (PLEG): container finished" podID="a0289180-1fc0-4d66-86ce-4e09fee7e0a9" containerID="18bfcabb32ad86abd878e7e940d829c71104702ca1658946a910f46d17f6b4d4" exitCode=0 Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.895043 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" event={"ID":"a0289180-1fc0-4d66-86ce-4e09fee7e0a9","Type":"ContainerDied","Data":"18bfcabb32ad86abd878e7e940d829c71104702ca1658946a910f46d17f6b4d4"} Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.897103 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" event={"ID":"6bcc761d-f5bd-476b-a8e5-25dcca68942e","Type":"ContainerStarted","Data":"ea379b46d1536fe40fe1df81aa6fea45d293a4a5bceca1514fbf5e3eb9bbffde"} Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.897126 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" event={"ID":"6bcc761d-f5bd-476b-a8e5-25dcca68942e","Type":"ContainerStarted","Data":"c7cbe2450a46799ddf70ee561ec2b6bba2c9ece1449805f8ba2c7ae6924806b7"} Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.897136 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" event={"ID":"6bcc761d-f5bd-476b-a8e5-25dcca68942e","Type":"ContainerStarted","Data":"3875cf5a19c3a7da14a86c7573fc86eabb110045219de24bc23bd2a620cd5f27"} Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.972379 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:07 crc kubenswrapper[4759]: E1125 19:36:07.972722 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.472703766 +0000 UTC m=+159.058651467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:07 crc kubenswrapper[4759]: I1125 19:36:07.991540 4759 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.012654 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wrlnm" podStartSLOduration=10.012639594 podStartE2EDuration="10.012639594s" podCreationTimestamp="2025-11-25 19:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:08.009707173 +0000 UTC m=+158.595654874" watchObservedRunningTime="2025-11-25 19:36:08.012639594 +0000 UTC m=+158.598587295" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.074540 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: E1125 19:36:08.076988 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.576972445 +0000 UTC m=+159.162920146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.176111 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:08 crc kubenswrapper[4759]: E1125 19:36:08.176443 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.676427324 +0000 UTC m=+159.262375025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.277921 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: E1125 19:36:08.278205 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 19:36:08.778193855 +0000 UTC m=+159.364141556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-khxh2" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.307220 4759 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T19:36:07.99156972Z","Handler":null,"Name":""} Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.310498 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.311439 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.313106 4759 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.313142 4759 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.313623 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.328904 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.378974 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.417944 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.480006 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.480078 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.480155 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6rn\" (UniqueName: \"kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.480191 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.510162 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.511075 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.513038 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.528062 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.558905 4759 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.558965 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.581772 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.581820 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.581871 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6rn\" (UniqueName: \"kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.582553 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.582860 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.590859 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-khxh2\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.610362 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6rn\" (UniqueName: \"kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn\") pod \"community-operators-nz425\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.632090 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.682569 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.683080 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxkst\" (UniqueName: \"kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.683200 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.704953 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.705876 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.774640 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.778578 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.787552 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.787646 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.787716 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxkst\" (UniqueName: \"kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.796974 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.797076 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.810308 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxkst\" (UniqueName: \"kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst\") pod \"certified-operators-l9lhq\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.821812 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.869114 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:08 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:08 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:08 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.869185 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.892101 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.892179 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.892218 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgrjz\" (UniqueName: \"kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.905377 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.912681 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.914143 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:08 crc kubenswrapper[4759]: I1125 19:36:08.914430 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.001647 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgrjz\" (UniqueName: \"kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.002844 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.003371 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.003460 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.003728 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.032422 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgrjz\" (UniqueName: \"kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz\") pod \"community-operators-lk6nr\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.074229 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.104826 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.104939 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.105028 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwdhn\" (UniqueName: \"kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.122515 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.185292 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.205910 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.205954 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.205996 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwdhn\" (UniqueName: \"kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.206529 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.206737 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.235306 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwdhn\" (UniqueName: \"kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn\") pod \"certified-operators-j62pl\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.235657 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.315860 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") pod \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.315906 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume\") pod \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.315950 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h26sc\" (UniqueName: \"kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc\") pod \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\" (UID: \"a0289180-1fc0-4d66-86ce-4e09fee7e0a9\") " Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.316768 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume" (OuterVolumeSpecName: "config-volume") pod "a0289180-1fc0-4d66-86ce-4e09fee7e0a9" (UID: "a0289180-1fc0-4d66-86ce-4e09fee7e0a9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.328071 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a0289180-1fc0-4d66-86ce-4e09fee7e0a9" (UID: "a0289180-1fc0-4d66-86ce-4e09fee7e0a9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.330328 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc" (OuterVolumeSpecName: "kube-api-access-h26sc") pod "a0289180-1fc0-4d66-86ce-4e09fee7e0a9" (UID: "a0289180-1fc0-4d66-86ce-4e09fee7e0a9"). InnerVolumeSpecName "kube-api-access-h26sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.379123 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.417149 4759 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.417179 4759 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.417189 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h26sc\" (UniqueName: \"kubernetes.io/projected/a0289180-1fc0-4d66-86ce-4e09fee7e0a9-kube-api-access-h26sc\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.529086 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.714293 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.849437 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:09 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:09 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:09 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.849593 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.941153 4759 generic.go:334] "Generic (PLEG): container finished" podID="401e8bc0-19b1-440e-8062-298d61c12e37" containerID="1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069" exitCode=0 Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.941212 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerDied","Data":"1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.941268 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerStarted","Data":"7c17d178803fcaa7a9838f1be9d8879f82b6aefa7464dd5007ad4a5cb4b80653"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.943698 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.944219 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" event={"ID":"4a6c7e51-7e1e-46d7-9d75-758019749635","Type":"ContainerStarted","Data":"65ac357a459134b8a52666bf82ec6b4aa82597dfeb0d37358f2b78e54919d37c"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.944253 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" event={"ID":"4a6c7e51-7e1e-46d7-9d75-758019749635","Type":"ContainerStarted","Data":"9058f6f9b3fe29d79dc8e21e974de822f495eb50afc689b93add91b455070cc8"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.944338 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.945510 4759 generic.go:334] "Generic (PLEG): container finished" podID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerID="bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8" exitCode=0 Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.945575 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerDied","Data":"bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.945598 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerStarted","Data":"87c5cd3110ba1e97a2a7a37c24e68286e95e9727aa1d4a18e2e0a82f5848cc16"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.950598 4759 generic.go:334] "Generic (PLEG): container finished" podID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerID="613ae1f8df12128c2f4f73baebacb75490d3cac7f47c11366174e3d751873781" exitCode=0 Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.950674 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerDied","Data":"613ae1f8df12128c2f4f73baebacb75490d3cac7f47c11366174e3d751873781"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.951829 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerStarted","Data":"1308f18c4147850715fec36545f3315d47ddab8ffddfd01b0eeae06ddb6fb1f4"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.956249 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" event={"ID":"a0289180-1fc0-4d66-86ce-4e09fee7e0a9","Type":"ContainerDied","Data":"71ddbe02c49e9c3fa93e4da8b9f3522691d054b36a5c977b523344848dc402a3"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.956290 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71ddbe02c49e9c3fa93e4da8b9f3522691d054b36a5c977b523344848dc402a3" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.956263 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401650-lfdkx" Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.962737 4759 generic.go:334] "Generic (PLEG): container finished" podID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerID="d53890e421115448a9a299b14ca963d3d3b5c5c014c49fad46ecf81e06a6923d" exitCode=0 Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.962793 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerDied","Data":"d53890e421115448a9a299b14ca963d3d3b5c5c014c49fad46ecf81e06a6923d"} Nov 25 19:36:09 crc kubenswrapper[4759]: I1125 19:36:09.962824 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerStarted","Data":"f1b2db4b1690b430e37450c10f0b26ce2d1ba9d89a544c0aa3fe15ac15730699"} Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.006652 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" podStartSLOduration=139.006633102 podStartE2EDuration="2m19.006633102s" podCreationTimestamp="2025-11-25 19:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:10.005413219 +0000 UTC m=+160.591360930" watchObservedRunningTime="2025-11-25 19:36:10.006633102 +0000 UTC m=+160.592580803" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.116819 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.506443 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:36:10 crc kubenswrapper[4759]: E1125 19:36:10.506684 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0289180-1fc0-4d66-86ce-4e09fee7e0a9" containerName="collect-profiles" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.506700 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0289180-1fc0-4d66-86ce-4e09fee7e0a9" containerName="collect-profiles" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.506991 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0289180-1fc0-4d66-86ce-4e09fee7e0a9" containerName="collect-profiles" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.507808 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.509291 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.524650 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.535034 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcsg8\" (UniqueName: \"kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.535096 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.535115 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.636803 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.636853 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.636953 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcsg8\" (UniqueName: \"kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.637434 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.637466 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.657573 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcsg8\" (UniqueName: \"kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8\") pod \"redhat-marketplace-wfh5l\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.825176 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.848700 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:10 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:10 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:10 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.848755 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.905152 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.906047 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:10 crc kubenswrapper[4759]: I1125 19:36:10.921082 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.045104 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.045417 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.045526 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r26xc\" (UniqueName: \"kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.134317 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.135722 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.140355 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.140541 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.141875 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.146121 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.146158 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.146290 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r26xc\" (UniqueName: \"kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.147568 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.147957 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.168083 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.170343 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r26xc\" (UniqueName: \"kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc\") pod \"redhat-marketplace-dntgx\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.235033 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.247799 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.247862 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.287643 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.287745 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.289499 4759 patch_prober.go:28] interesting pod/console-f9d7485db-4szh4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.289584 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4szh4" podUID="32444b14-e189-4a50-8e25-ed6323de94c7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.348801 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.348854 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.348982 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.365249 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.466815 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.512319 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.513865 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.519916 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.521353 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.591134 4759 patch_prober.go:28] interesting pod/downloads-7954f5f757-x8rv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.591230 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-x8rv5" podUID="4e9ac029-8117-49ff-be41-19f18e9f621a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.591665 4759 patch_prober.go:28] interesting pod/downloads-7954f5f757-x8rv5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.591708 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-x8rv5" podUID="4e9ac029-8117-49ff-be41-19f18e9f621a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.655131 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl5p2\" (UniqueName: \"kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.655193 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.655245 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.724590 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:11 crc kubenswrapper[4759]: W1125 19:36:11.751334 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb909b865_945e_42d2_a394_df09627d4931.slice/crio-aab951be236e157705103f0384717aac831795c61fa7e5ff3ec5ebc48bb47151 WatchSource:0}: Error finding container aab951be236e157705103f0384717aac831795c61fa7e5ff3ec5ebc48bb47151: Status 404 returned error can't find the container with id aab951be236e157705103f0384717aac831795c61fa7e5ff3ec5ebc48bb47151 Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.756301 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.756634 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl5p2\" (UniqueName: \"kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.756689 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.758216 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.758691 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.763516 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.801062 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl5p2\" (UniqueName: \"kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2\") pod \"redhat-operators-6zx9x\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.802803 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 19:36:11 crc kubenswrapper[4759]: W1125 19:36:11.822910 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5c02ae30_ebca_48ba_9b57_304fb01a6d1b.slice/crio-e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7 WatchSource:0}: Error finding container e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7: Status 404 returned error can't find the container with id e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7 Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.847125 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.851606 4759 patch_prober.go:28] interesting pod/router-default-5444994796-k7nmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 19:36:11 crc kubenswrapper[4759]: [-]has-synced failed: reason withheld Nov 25 19:36:11 crc kubenswrapper[4759]: [+]process-running ok Nov 25 19:36:11 crc kubenswrapper[4759]: healthz check failed Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.851651 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k7nmp" podUID="2c0b0690-6f47-4f2b-bb68-fd7e413ab7e5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.880970 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.914221 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.915281 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.919101 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.934524 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.939581 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-vzt92" Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.996369 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5c02ae30-ebca-48ba-9b57-304fb01a6d1b","Type":"ContainerStarted","Data":"e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7"} Nov 25 19:36:11 crc kubenswrapper[4759]: I1125 19:36:11.997945 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerStarted","Data":"aab951be236e157705103f0384717aac831795c61fa7e5ff3ec5ebc48bb47151"} Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:11.999526 4759 generic.go:334] "Generic (PLEG): container finished" podID="a685e63a-c301-4410-84f7-6825651ad294" containerID="d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c" exitCode=0 Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.000491 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerDied","Data":"d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c"} Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.000508 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerStarted","Data":"c1b7c182aaffc478ae9f103c9856fc7ee741bd0ce27403034ad4c98a46df3930"} Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.060315 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbqr5\" (UniqueName: \"kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.060376 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.060421 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.161222 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.161369 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbqr5\" (UniqueName: \"kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.161409 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.162391 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.163499 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.192402 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbqr5\" (UniqueName: \"kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5\") pod \"redhat-operators-xlnzt\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.240945 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.311341 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:36:12 crc kubenswrapper[4759]: W1125 19:36:12.322229 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3eb78bd8_c3b4_4daa_a818_07501af8098f.slice/crio-096f077670c195bb8b992069ffe3871b61d623e927a8f263abdbd3ed66f2c59b WatchSource:0}: Error finding container 096f077670c195bb8b992069ffe3871b61d623e927a8f263abdbd3ed66f2c59b: Status 404 returned error can't find the container with id 096f077670c195bb8b992069ffe3871b61d623e927a8f263abdbd3ed66f2c59b Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.529019 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:12 crc kubenswrapper[4759]: W1125 19:36:12.548326 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcee13927_268e_4a4b_95a3_d1fe5c9cf204.slice/crio-22db1ffb717b2a60046883653bb7829ce36939cfa9d7168ad7fa9a2ee9333e33 WatchSource:0}: Error finding container 22db1ffb717b2a60046883653bb7829ce36939cfa9d7168ad7fa9a2ee9333e33: Status 404 returned error can't find the container with id 22db1ffb717b2a60046883653bb7829ce36939cfa9d7168ad7fa9a2ee9333e33 Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.856021 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:12 crc kubenswrapper[4759]: I1125 19:36:12.863703 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-k7nmp" Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.009620 4759 generic.go:334] "Generic (PLEG): container finished" podID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerID="9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489" exitCode=0 Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.009712 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerDied","Data":"9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.009787 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerStarted","Data":"096f077670c195bb8b992069ffe3871b61d623e927a8f263abdbd3ed66f2c59b"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.024912 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5c02ae30-ebca-48ba-9b57-304fb01a6d1b","Type":"ContainerStarted","Data":"26c4a0457628a4d4431c26725e2d4d5d39365a3ac6e0d9d379798d3d0d098dec"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.027513 4759 generic.go:334] "Generic (PLEG): container finished" podID="b909b865-945e-42d2-a394-df09627d4931" containerID="9436a93c111ea3b119e46428203696e3350857c3ff66428f0fcfc62dd548e192" exitCode=0 Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.027555 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerDied","Data":"9436a93c111ea3b119e46428203696e3350857c3ff66428f0fcfc62dd548e192"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.041189 4759 generic.go:334] "Generic (PLEG): container finished" podID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerID="939cf094dcd9f0475b91f90d427c63d5da64d7eabf04d381caf6eb0adec6395f" exitCode=0 Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.041298 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerDied","Data":"939cf094dcd9f0475b91f90d427c63d5da64d7eabf04d381caf6eb0adec6395f"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.041363 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerStarted","Data":"22db1ffb717b2a60046883653bb7829ce36939cfa9d7168ad7fa9a2ee9333e33"} Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.481065 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.489938 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e41227a-a128-4f1e-a21c-2242cea52970-metrics-certs\") pod \"network-metrics-daemon-k5q6v\" (UID: \"8e41227a-a128-4f1e-a21c-2242cea52970\") " pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:36:13 crc kubenswrapper[4759]: I1125 19:36:13.638058 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k5q6v" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.003047 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k5q6v"] Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.064293 4759 generic.go:334] "Generic (PLEG): container finished" podID="5c02ae30-ebca-48ba-9b57-304fb01a6d1b" containerID="26c4a0457628a4d4431c26725e2d4d5d39365a3ac6e0d9d379798d3d0d098dec" exitCode=0 Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.064395 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5c02ae30-ebca-48ba-9b57-304fb01a6d1b","Type":"ContainerDied","Data":"26c4a0457628a4d4431c26725e2d4d5d39365a3ac6e0d9d379798d3d0d098dec"} Nov 25 19:36:14 crc kubenswrapper[4759]: W1125 19:36:14.075355 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e41227a_a128_4f1e_a21c_2242cea52970.slice/crio-92b1347ddd7ba3bfd55a0a972bc62812700649918831da743e6dd9455fa3ae2f WatchSource:0}: Error finding container 92b1347ddd7ba3bfd55a0a972bc62812700649918831da743e6dd9455fa3ae2f: Status 404 returned error can't find the container with id 92b1347ddd7ba3bfd55a0a972bc62812700649918831da743e6dd9455fa3ae2f Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.137080 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mqn2l" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.414939 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.517246 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir\") pod \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.517330 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access\") pod \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\" (UID: \"5c02ae30-ebca-48ba-9b57-304fb01a6d1b\") " Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.517392 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5c02ae30-ebca-48ba-9b57-304fb01a6d1b" (UID: "5c02ae30-ebca-48ba-9b57-304fb01a6d1b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.517671 4759 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.530664 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5c02ae30-ebca-48ba-9b57-304fb01a6d1b" (UID: "5c02ae30-ebca-48ba-9b57-304fb01a6d1b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:14 crc kubenswrapper[4759]: I1125 19:36:14.618659 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5c02ae30-ebca-48ba-9b57-304fb01a6d1b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.078606 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" event={"ID":"8e41227a-a128-4f1e-a21c-2242cea52970","Type":"ContainerStarted","Data":"92b1347ddd7ba3bfd55a0a972bc62812700649918831da743e6dd9455fa3ae2f"} Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.081679 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5c02ae30-ebca-48ba-9b57-304fb01a6d1b","Type":"ContainerDied","Data":"e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7"} Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.081727 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.081744 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4eae16dab55a838864427ff9a59255c6a8e745b3b1e58195719b3da6e0f5ea7" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.239165 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 19:36:15 crc kubenswrapper[4759]: E1125 19:36:15.241437 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c02ae30-ebca-48ba-9b57-304fb01a6d1b" containerName="pruner" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.243025 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c02ae30-ebca-48ba-9b57-304fb01a6d1b" containerName="pruner" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.243319 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c02ae30-ebca-48ba-9b57-304fb01a6d1b" containerName="pruner" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.243919 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.258930 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.260526 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.261431 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.332179 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.332250 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.433366 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.433472 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.433512 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.449539 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:15 crc kubenswrapper[4759]: I1125 19:36:15.567823 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:16 crc kubenswrapper[4759]: I1125 19:36:16.098370 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" event={"ID":"8e41227a-a128-4f1e-a21c-2242cea52970","Type":"ContainerStarted","Data":"c3d3c0d41785bc67f60569be05609e0daf67c79123792d4e8aec06925027430d"} Nov 25 19:36:16 crc kubenswrapper[4759]: I1125 19:36:16.116995 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 19:36:17 crc kubenswrapper[4759]: I1125 19:36:17.107119 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bb4e8284-eaf5-46dc-abac-b48d50da6f14","Type":"ContainerStarted","Data":"d6c38e7bd1660bc9b0df61dd6b2e0c2bfd2170527c242555800429e87004c4d7"} Nov 25 19:36:17 crc kubenswrapper[4759]: I1125 19:36:17.110145 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k5q6v" event={"ID":"8e41227a-a128-4f1e-a21c-2242cea52970","Type":"ContainerStarted","Data":"8b29274cc9c5889a487b440a547dad440890cbc9f67cbee77c877800c8ba8d0c"} Nov 25 19:36:17 crc kubenswrapper[4759]: I1125 19:36:17.132635 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-k5q6v" podStartSLOduration=147.13261424 podStartE2EDuration="2m27.13261424s" podCreationTimestamp="2025-11-25 19:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:17.128313433 +0000 UTC m=+167.714261144" watchObservedRunningTime="2025-11-25 19:36:17.13261424 +0000 UTC m=+167.718561951" Nov 25 19:36:21 crc kubenswrapper[4759]: I1125 19:36:21.292849 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:21 crc kubenswrapper[4759]: I1125 19:36:21.297600 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4szh4" Nov 25 19:36:21 crc kubenswrapper[4759]: I1125 19:36:21.596624 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-x8rv5" Nov 25 19:36:24 crc kubenswrapper[4759]: I1125 19:36:24.895998 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:36:24 crc kubenswrapper[4759]: I1125 19:36:24.896330 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:36:28 crc kubenswrapper[4759]: I1125 19:36:28.474635 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 19:36:28 crc kubenswrapper[4759]: I1125 19:36:28.782510 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:36:36 crc kubenswrapper[4759]: E1125 19:36:36.619687 4759 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 19:36:36 crc kubenswrapper[4759]: E1125 19:36:36.620458 4759 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fgrjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lk6nr_openshift-marketplace(62dce71f-f0d6-4044-84fd-bc456278eab9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 19:36:36 crc kubenswrapper[4759]: E1125 19:36:36.622056 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lk6nr" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" Nov 25 19:36:40 crc kubenswrapper[4759]: E1125 19:36:40.199782 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lk6nr" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" Nov 25 19:36:40 crc kubenswrapper[4759]: E1125 19:36:40.332652 4759 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 19:36:40 crc kubenswrapper[4759]: E1125 19:36:40.332772 4759 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fcsg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wfh5l_openshift-marketplace(a685e63a-c301-4410-84f7-6825651ad294): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 19:36:40 crc kubenswrapper[4759]: E1125 19:36:40.333846 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wfh5l" podUID="a685e63a-c301-4410-84f7-6825651ad294" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.636891 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wfh5l" podUID="a685e63a-c301-4410-84f7-6825651ad294" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.723884 4759 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.724518 4759 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gxkst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-l9lhq_openshift-marketplace(d0ded011-6d23-4501-a7ea-f90ed6696cdd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.726655 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-l9lhq" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.753635 4759 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.753762 4759 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwdhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-j62pl_openshift-marketplace(967ed015-8b0e-4c14-922f-2df7857bdcb6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 19:36:41 crc kubenswrapper[4759]: E1125 19:36:41.755189 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-j62pl" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.349572 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerStarted","Data":"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a"} Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.352727 4759 generic.go:334] "Generic (PLEG): container finished" podID="401e8bc0-19b1-440e-8062-298d61c12e37" containerID="2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6" exitCode=0 Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.352794 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerDied","Data":"2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6"} Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.355323 4759 generic.go:334] "Generic (PLEG): container finished" podID="b909b865-945e-42d2-a394-df09627d4931" containerID="6a0bbed99b13f185dd0776371788fba31ee79e2d2e40b69e8597077a01d3cccc" exitCode=0 Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.355386 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerDied","Data":"6a0bbed99b13f185dd0776371788fba31ee79e2d2e40b69e8597077a01d3cccc"} Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.358562 4759 generic.go:334] "Generic (PLEG): container finished" podID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerID="9fb48dff3cabd8dc0d909cbb0084209f4658450e57ff07565f42c838cbecce23" exitCode=0 Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.358634 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerDied","Data":"9fb48dff3cabd8dc0d909cbb0084209f4658450e57ff07565f42c838cbecce23"} Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.359933 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bb4e8284-eaf5-46dc-abac-b48d50da6f14","Type":"ContainerStarted","Data":"3d6e6aae9874274cbbb520c41f128e2092cc9c76e8d60aa0a420d6bd6150e624"} Nov 25 19:36:42 crc kubenswrapper[4759]: E1125 19:36:42.361895 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-l9lhq" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" Nov 25 19:36:42 crc kubenswrapper[4759]: E1125 19:36:42.362943 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-j62pl" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.536774 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7skhl" Nov 25 19:36:42 crc kubenswrapper[4759]: I1125 19:36:42.556534 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=27.556515818 podStartE2EDuration="27.556515818s" podCreationTimestamp="2025-11-25 19:36:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:42.49563974 +0000 UTC m=+193.081587441" watchObservedRunningTime="2025-11-25 19:36:42.556515818 +0000 UTC m=+193.142463519" Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.366050 4759 generic.go:334] "Generic (PLEG): container finished" podID="bb4e8284-eaf5-46dc-abac-b48d50da6f14" containerID="3d6e6aae9874274cbbb520c41f128e2092cc9c76e8d60aa0a420d6bd6150e624" exitCode=0 Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.366322 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bb4e8284-eaf5-46dc-abac-b48d50da6f14","Type":"ContainerDied","Data":"3d6e6aae9874274cbbb520c41f128e2092cc9c76e8d60aa0a420d6bd6150e624"} Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.368432 4759 generic.go:334] "Generic (PLEG): container finished" podID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerID="fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a" exitCode=0 Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.368682 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerDied","Data":"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a"} Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.372244 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerStarted","Data":"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983"} Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.374299 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerStarted","Data":"f0fdbf530d2e47e8c02bc1cd2641cf8be2aa1005e6b1c3d731eb6d6c2db75c87"} Nov 25 19:36:43 crc kubenswrapper[4759]: I1125 19:36:43.414956 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nz425" podStartSLOduration=2.361165487 podStartE2EDuration="35.414937934s" podCreationTimestamp="2025-11-25 19:36:08 +0000 UTC" firstStartedPulling="2025-11-25 19:36:09.943466972 +0000 UTC m=+160.529414673" lastFinishedPulling="2025-11-25 19:36:42.997239419 +0000 UTC m=+193.583187120" observedRunningTime="2025-11-25 19:36:43.412893678 +0000 UTC m=+193.998841379" watchObservedRunningTime="2025-11-25 19:36:43.414937934 +0000 UTC m=+194.000885655" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.385746 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerStarted","Data":"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86"} Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.388440 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerStarted","Data":"f664d4f5fa2ecb1ab4432a809cc9a0a0e81fb92f650d1922095fddbef1b86e4f"} Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.401266 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dntgx" podStartSLOduration=4.303830609 podStartE2EDuration="34.401248922s" podCreationTimestamp="2025-11-25 19:36:10 +0000 UTC" firstStartedPulling="2025-11-25 19:36:13.032351715 +0000 UTC m=+163.618299416" lastFinishedPulling="2025-11-25 19:36:43.129770028 +0000 UTC m=+193.715717729" observedRunningTime="2025-11-25 19:36:43.43096254 +0000 UTC m=+194.016910271" watchObservedRunningTime="2025-11-25 19:36:44.401248922 +0000 UTC m=+194.987196623" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.402161 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6zx9x" podStartSLOduration=2.589257391 podStartE2EDuration="33.402156376s" podCreationTimestamp="2025-11-25 19:36:11 +0000 UTC" firstStartedPulling="2025-11-25 19:36:13.011782436 +0000 UTC m=+163.597730137" lastFinishedPulling="2025-11-25 19:36:43.824681421 +0000 UTC m=+194.410629122" observedRunningTime="2025-11-25 19:36:44.400573804 +0000 UTC m=+194.986521505" watchObservedRunningTime="2025-11-25 19:36:44.402156376 +0000 UTC m=+194.988104077" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.415396 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xlnzt" podStartSLOduration=3.251778262 podStartE2EDuration="33.415376266s" podCreationTimestamp="2025-11-25 19:36:11 +0000 UTC" firstStartedPulling="2025-11-25 19:36:13.044720653 +0000 UTC m=+163.630668354" lastFinishedPulling="2025-11-25 19:36:43.208318657 +0000 UTC m=+193.794266358" observedRunningTime="2025-11-25 19:36:44.414369559 +0000 UTC m=+195.000317260" watchObservedRunningTime="2025-11-25 19:36:44.415376266 +0000 UTC m=+195.001323957" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.792098 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.910021 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir\") pod \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.910144 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb4e8284-eaf5-46dc-abac-b48d50da6f14" (UID: "bb4e8284-eaf5-46dc-abac-b48d50da6f14"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.910173 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access\") pod \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\" (UID: \"bb4e8284-eaf5-46dc-abac-b48d50da6f14\") " Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.910423 4759 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:44 crc kubenswrapper[4759]: I1125 19:36:44.914695 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb4e8284-eaf5-46dc-abac-b48d50da6f14" (UID: "bb4e8284-eaf5-46dc-abac-b48d50da6f14"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:45 crc kubenswrapper[4759]: I1125 19:36:45.012342 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb4e8284-eaf5-46dc-abac-b48d50da6f14-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:45 crc kubenswrapper[4759]: I1125 19:36:45.395310 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bb4e8284-eaf5-46dc-abac-b48d50da6f14","Type":"ContainerDied","Data":"d6c38e7bd1660bc9b0df61dd6b2e0c2bfd2170527c242555800429e87004c4d7"} Nov 25 19:36:45 crc kubenswrapper[4759]: I1125 19:36:45.395359 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6c38e7bd1660bc9b0df61dd6b2e0c2bfd2170527c242555800429e87004c4d7" Nov 25 19:36:45 crc kubenswrapper[4759]: I1125 19:36:45.395405 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 19:36:48 crc kubenswrapper[4759]: I1125 19:36:48.632892 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:48 crc kubenswrapper[4759]: I1125 19:36:48.633170 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:49 crc kubenswrapper[4759]: I1125 19:36:49.023402 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:49 crc kubenswrapper[4759]: I1125 19:36:49.452298 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.235971 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.237047 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.282191 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.461590 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.881806 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.882177 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:51 crc kubenswrapper[4759]: I1125 19:36:51.921196 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.119274 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.242895 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.243518 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.281003 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.463300 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.467156 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.638851 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 19:36:52 crc kubenswrapper[4759]: E1125 19:36:52.639055 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4e8284-eaf5-46dc-abac-b48d50da6f14" containerName="pruner" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.639066 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4e8284-eaf5-46dc-abac-b48d50da6f14" containerName="pruner" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.639152 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4e8284-eaf5-46dc-abac-b48d50da6f14" containerName="pruner" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.639517 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.641094 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.641319 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.650228 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.805121 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.805308 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.906834 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.906902 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.906983 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.928708 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:52 crc kubenswrapper[4759]: I1125 19:36:52.955522 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:53 crc kubenswrapper[4759]: I1125 19:36:53.354781 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 19:36:53 crc kubenswrapper[4759]: W1125 19:36:53.365397 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poddc7ea4bf_0efa_4fe6_8035_dc6ca6757a29.slice/crio-59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f WatchSource:0}: Error finding container 59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f: Status 404 returned error can't find the container with id 59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f Nov 25 19:36:53 crc kubenswrapper[4759]: I1125 19:36:53.430992 4759 generic.go:334] "Generic (PLEG): container finished" podID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerID="992f25b89add659810e7af2e9a548d0caa267305747be7221e0148e19c87b251" exitCode=0 Nov 25 19:36:53 crc kubenswrapper[4759]: I1125 19:36:53.431103 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerDied","Data":"992f25b89add659810e7af2e9a548d0caa267305747be7221e0148e19c87b251"} Nov 25 19:36:53 crc kubenswrapper[4759]: I1125 19:36:53.432933 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29","Type":"ContainerStarted","Data":"59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f"} Nov 25 19:36:53 crc kubenswrapper[4759]: I1125 19:36:53.433338 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dntgx" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="registry-server" containerID="cri-o://f0fdbf530d2e47e8c02bc1cd2641cf8be2aa1005e6b1c3d731eb6d6c2db75c87" gracePeriod=2 Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.308248 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.457825 4759 generic.go:334] "Generic (PLEG): container finished" podID="b909b865-945e-42d2-a394-df09627d4931" containerID="f0fdbf530d2e47e8c02bc1cd2641cf8be2aa1005e6b1c3d731eb6d6c2db75c87" exitCode=0 Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.457867 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerDied","Data":"f0fdbf530d2e47e8c02bc1cd2641cf8be2aa1005e6b1c3d731eb6d6c2db75c87"} Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.459952 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29","Type":"ContainerStarted","Data":"a7d03e151620b5405afdeb1c19c3a3709b4dcf5f59753ad6d9974514dbe1524d"} Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.484203 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.484084318 podStartE2EDuration="2.484084318s" podCreationTimestamp="2025-11-25 19:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:36:54.478027524 +0000 UTC m=+205.063975235" watchObservedRunningTime="2025-11-25 19:36:54.484084318 +0000 UTC m=+205.070032019" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.604046 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.756069 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities\") pod \"b909b865-945e-42d2-a394-df09627d4931\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.756186 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content\") pod \"b909b865-945e-42d2-a394-df09627d4931\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.756246 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r26xc\" (UniqueName: \"kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc\") pod \"b909b865-945e-42d2-a394-df09627d4931\" (UID: \"b909b865-945e-42d2-a394-df09627d4931\") " Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.757834 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities" (OuterVolumeSpecName: "utilities") pod "b909b865-945e-42d2-a394-df09627d4931" (UID: "b909b865-945e-42d2-a394-df09627d4931"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.761356 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc" (OuterVolumeSpecName: "kube-api-access-r26xc") pod "b909b865-945e-42d2-a394-df09627d4931" (UID: "b909b865-945e-42d2-a394-df09627d4931"). InnerVolumeSpecName "kube-api-access-r26xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.775399 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b909b865-945e-42d2-a394-df09627d4931" (UID: "b909b865-945e-42d2-a394-df09627d4931"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.857357 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r26xc\" (UniqueName: \"kubernetes.io/projected/b909b865-945e-42d2-a394-df09627d4931-kube-api-access-r26xc\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.857401 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.857414 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b909b865-945e-42d2-a394-df09627d4931-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.895852 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.896177 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.896227 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.896854 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:36:54 crc kubenswrapper[4759]: I1125 19:36:54.896954 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf" gracePeriod=600 Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.466967 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dntgx" event={"ID":"b909b865-945e-42d2-a394-df09627d4931","Type":"ContainerDied","Data":"aab951be236e157705103f0384717aac831795c61fa7e5ff3ec5ebc48bb47151"} Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.467003 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dntgx" Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.467027 4759 scope.go:117] "RemoveContainer" containerID="f0fdbf530d2e47e8c02bc1cd2641cf8be2aa1005e6b1c3d731eb6d6c2db75c87" Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.468603 4759 generic.go:334] "Generic (PLEG): container finished" podID="dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" containerID="a7d03e151620b5405afdeb1c19c3a3709b4dcf5f59753ad6d9974514dbe1524d" exitCode=0 Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.468708 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29","Type":"ContainerDied","Data":"a7d03e151620b5405afdeb1c19c3a3709b4dcf5f59753ad6d9974514dbe1524d"} Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.468836 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xlnzt" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="registry-server" containerID="cri-o://f664d4f5fa2ecb1ab4432a809cc9a0a0e81fb92f650d1922095fddbef1b86e4f" gracePeriod=2 Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.503787 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:55 crc kubenswrapper[4759]: I1125 19:36:55.504853 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dntgx"] Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.121368 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b909b865-945e-42d2-a394-df09627d4931" path="/var/lib/kubelet/pods/b909b865-945e-42d2-a394-df09627d4931/volumes" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.208465 4759 scope.go:117] "RemoveContainer" containerID="6a0bbed99b13f185dd0776371788fba31ee79e2d2e40b69e8597077a01d3cccc" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.409986 4759 scope.go:117] "RemoveContainer" containerID="9436a93c111ea3b119e46428203696e3350857c3ff66428f0fcfc62dd548e192" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.475141 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf" exitCode=0 Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.475204 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf"} Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.478422 4759 generic.go:334] "Generic (PLEG): container finished" podID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerID="f664d4f5fa2ecb1ab4432a809cc9a0a0e81fb92f650d1922095fddbef1b86e4f" exitCode=0 Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.478494 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerDied","Data":"f664d4f5fa2ecb1ab4432a809cc9a0a0e81fb92f650d1922095fddbef1b86e4f"} Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.854507 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.980628 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir\") pod \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.980707 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access\") pod \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\" (UID: \"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29\") " Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.980738 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" (UID: "dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.980943 4759 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:56 crc kubenswrapper[4759]: I1125 19:36:56.986636 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" (UID: "dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.082182 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.469741 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.488433 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29","Type":"ContainerDied","Data":"59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f"} Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.488783 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59162968cbf6e0f455e7985389b32699e7461e829007507b3fa2ba39f3ec1b8f" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.488462 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.491810 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlnzt" event={"ID":"cee13927-268e-4a4b-95a3-d1fe5c9cf204","Type":"ContainerDied","Data":"22db1ffb717b2a60046883653bb7829ce36939cfa9d7168ad7fa9a2ee9333e33"} Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.491831 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlnzt" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.491881 4759 scope.go:117] "RemoveContainer" containerID="f664d4f5fa2ecb1ab4432a809cc9a0a0e81fb92f650d1922095fddbef1b86e4f" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.516370 4759 scope.go:117] "RemoveContainer" containerID="9fb48dff3cabd8dc0d909cbb0084209f4658450e57ff07565f42c838cbecce23" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.533863 4759 scope.go:117] "RemoveContainer" containerID="939cf094dcd9f0475b91f90d427c63d5da64d7eabf04d381caf6eb0adec6395f" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.586901 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbqr5\" (UniqueName: \"kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5\") pod \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.586981 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content\") pod \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.587023 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities\") pod \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\" (UID: \"cee13927-268e-4a4b-95a3-d1fe5c9cf204\") " Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.587869 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities" (OuterVolumeSpecName: "utilities") pod "cee13927-268e-4a4b-95a3-d1fe5c9cf204" (UID: "cee13927-268e-4a4b-95a3-d1fe5c9cf204"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.591033 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5" (OuterVolumeSpecName: "kube-api-access-nbqr5") pod "cee13927-268e-4a4b-95a3-d1fe5c9cf204" (UID: "cee13927-268e-4a4b-95a3-d1fe5c9cf204"). InnerVolumeSpecName "kube-api-access-nbqr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.688775 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbqr5\" (UniqueName: \"kubernetes.io/projected/cee13927-268e-4a4b-95a3-d1fe5c9cf204-kube-api-access-nbqr5\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.688806 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.894911 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cee13927-268e-4a4b-95a3-d1fe5c9cf204" (UID: "cee13927-268e-4a4b-95a3-d1fe5c9cf204"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:36:57 crc kubenswrapper[4759]: I1125 19:36:57.991870 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee13927-268e-4a4b-95a3-d1fe5c9cf204-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.119646 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.122438 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xlnzt"] Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.516697 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab"} Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.518796 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerStarted","Data":"4f2e40a5b56fcf73d513bc80cc303a26d941ef4e222248207b75fb10c817373d"} Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.539159 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lk6nr" podStartSLOduration=3.126706495 podStartE2EDuration="50.539139208s" podCreationTimestamp="2025-11-25 19:36:08 +0000 UTC" firstStartedPulling="2025-11-25 19:36:09.964390672 +0000 UTC m=+160.550338393" lastFinishedPulling="2025-11-25 19:36:57.376823405 +0000 UTC m=+207.962771106" observedRunningTime="2025-11-25 19:36:58.536516345 +0000 UTC m=+209.122464046" watchObservedRunningTime="2025-11-25 19:36:58.539139208 +0000 UTC m=+209.125086909" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634567 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634803 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="extract-content" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634824 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="extract-content" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634839 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" containerName="pruner" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634850 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" containerName="pruner" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634860 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634868 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634879 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="extract-utilities" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634886 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="extract-utilities" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634895 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="extract-content" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634901 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="extract-content" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634911 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="extract-utilities" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634918 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="extract-utilities" Nov 25 19:36:58 crc kubenswrapper[4759]: E1125 19:36:58.634925 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.634933 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.635049 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7ea4bf-0efa-4fe6-8035-dc6ca6757a29" containerName="pruner" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.635062 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.635077 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b909b865-945e-42d2-a394-df09627d4931" containerName="registry-server" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.635803 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.640623 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.642362 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.647255 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.801532 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.801581 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.801649 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.902922 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.902976 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.903033 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.903064 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.903130 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.919872 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access\") pod \"installer-9-crc\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:58 crc kubenswrapper[4759]: I1125 19:36:58.972938 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:36:59 crc kubenswrapper[4759]: I1125 19:36:59.123302 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:59 crc kubenswrapper[4759]: I1125 19:36:59.123378 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:36:59 crc kubenswrapper[4759]: I1125 19:36:59.579820 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.115172 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee13927-268e-4a4b-95a3-d1fe5c9cf204" path="/var/lib/kubelet/pods/cee13927-268e-4a4b-95a3-d1fe5c9cf204/volumes" Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.164641 4759 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-lk6nr" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="registry-server" probeResult="failure" output=< Nov 25 19:37:00 crc kubenswrapper[4759]: timeout: failed to connect service ":50051" within 1s Nov 25 19:37:00 crc kubenswrapper[4759]: > Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.529098 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"254f97f7-20ca-4def-a04d-6ef872a5eb70","Type":"ContainerStarted","Data":"2559372e841f66a750a4a84f8002f6724835e725b9430b9cde0b6fea0d272e8c"} Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.529373 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"254f97f7-20ca-4def-a04d-6ef872a5eb70","Type":"ContainerStarted","Data":"e70eb6e6a3113750ccda4e9fb3af44719c4b03c38a84d333b3a8fd8721a1bbde"} Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.532073 4759 generic.go:334] "Generic (PLEG): container finished" podID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerID="598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073" exitCode=0 Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.532141 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerDied","Data":"598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073"} Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.534546 4759 generic.go:334] "Generic (PLEG): container finished" podID="a685e63a-c301-4410-84f7-6825651ad294" containerID="3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3" exitCode=0 Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.534614 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerDied","Data":"3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3"} Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.536839 4759 generic.go:334] "Generic (PLEG): container finished" podID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerID="d7273dbb4cb3477872aa890059f8f5c16ca1d8523150fae7e05b4e4ba3ddbc5a" exitCode=0 Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.536870 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerDied","Data":"d7273dbb4cb3477872aa890059f8f5c16ca1d8523150fae7e05b4e4ba3ddbc5a"} Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.569192 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.569171411 podStartE2EDuration="2.569171411s" podCreationTimestamp="2025-11-25 19:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:37:00.543329316 +0000 UTC m=+211.129277047" watchObservedRunningTime="2025-11-25 19:37:00.569171411 +0000 UTC m=+211.155119112" Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.992378 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:37:00 crc kubenswrapper[4759]: I1125 19:37:00.992701 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerName="controller-manager" containerID="cri-o://ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40" gracePeriod=30 Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.084398 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.084774 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerName="route-controller-manager" containerID="cri-o://bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3" gracePeriod=30 Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.370556 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.434488 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config\") pod \"a53e86aa-8a32-40af-a371-0055d170c5ce\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.434540 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert\") pod \"a53e86aa-8a32-40af-a371-0055d170c5ce\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.434568 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfmlf\" (UniqueName: \"kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf\") pod \"a53e86aa-8a32-40af-a371-0055d170c5ce\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.434594 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca\") pod \"a53e86aa-8a32-40af-a371-0055d170c5ce\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.434661 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles\") pod \"a53e86aa-8a32-40af-a371-0055d170c5ce\" (UID: \"a53e86aa-8a32-40af-a371-0055d170c5ce\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.436725 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config" (OuterVolumeSpecName: "config") pod "a53e86aa-8a32-40af-a371-0055d170c5ce" (UID: "a53e86aa-8a32-40af-a371-0055d170c5ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.438178 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a53e86aa-8a32-40af-a371-0055d170c5ce" (UID: "a53e86aa-8a32-40af-a371-0055d170c5ce"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.438396 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca" (OuterVolumeSpecName: "client-ca") pod "a53e86aa-8a32-40af-a371-0055d170c5ce" (UID: "a53e86aa-8a32-40af-a371-0055d170c5ce"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.442969 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf" (OuterVolumeSpecName: "kube-api-access-vfmlf") pod "a53e86aa-8a32-40af-a371-0055d170c5ce" (UID: "a53e86aa-8a32-40af-a371-0055d170c5ce"). InnerVolumeSpecName "kube-api-access-vfmlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.443417 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.443929 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a53e86aa-8a32-40af-a371-0055d170c5ce" (UID: "a53e86aa-8a32-40af-a371-0055d170c5ce"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.536630 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z86ng\" (UniqueName: \"kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng\") pod \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.536694 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert\") pod \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.536771 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config\") pod \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.536796 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca\") pod \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\" (UID: \"6f9089f5-e196-48ad-9d8c-a14544b6a96f\") " Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.536989 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537004 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a53e86aa-8a32-40af-a371-0055d170c5ce-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537012 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfmlf\" (UniqueName: \"kubernetes.io/projected/a53e86aa-8a32-40af-a371-0055d170c5ce-kube-api-access-vfmlf\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537025 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537034 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a53e86aa-8a32-40af-a371-0055d170c5ce-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537546 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca" (OuterVolumeSpecName: "client-ca") pod "6f9089f5-e196-48ad-9d8c-a14544b6a96f" (UID: "6f9089f5-e196-48ad-9d8c-a14544b6a96f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.537897 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config" (OuterVolumeSpecName: "config") pod "6f9089f5-e196-48ad-9d8c-a14544b6a96f" (UID: "6f9089f5-e196-48ad-9d8c-a14544b6a96f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.540149 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6f9089f5-e196-48ad-9d8c-a14544b6a96f" (UID: "6f9089f5-e196-48ad-9d8c-a14544b6a96f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.540247 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng" (OuterVolumeSpecName: "kube-api-access-z86ng") pod "6f9089f5-e196-48ad-9d8c-a14544b6a96f" (UID: "6f9089f5-e196-48ad-9d8c-a14544b6a96f"). InnerVolumeSpecName "kube-api-access-z86ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.543024 4759 generic.go:334] "Generic (PLEG): container finished" podID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerID="bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3" exitCode=0 Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.543074 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" event={"ID":"6f9089f5-e196-48ad-9d8c-a14544b6a96f","Type":"ContainerDied","Data":"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.543100 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" event={"ID":"6f9089f5-e196-48ad-9d8c-a14544b6a96f","Type":"ContainerDied","Data":"b8aaf6770879642414266e82093be4b9699f5c3adcdbf83d7e2b5f75414a93c6"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.543115 4759 scope.go:117] "RemoveContainer" containerID="bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.543207 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.549832 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerStarted","Data":"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.553324 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerStarted","Data":"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.556784 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerStarted","Data":"3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.562079 4759 scope.go:117] "RemoveContainer" containerID="bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3" Nov 25 19:37:01 crc kubenswrapper[4759]: E1125 19:37:01.562552 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3\": container with ID starting with bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3 not found: ID does not exist" containerID="bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.562588 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3"} err="failed to get container status \"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3\": rpc error: code = NotFound desc = could not find container \"bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3\": container with ID starting with bde2a2b6ad16c6d041943ac5ea66d6c28e601d418b5deca6ba4eeb8780d891f3 not found: ID does not exist" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.563125 4759 generic.go:334] "Generic (PLEG): container finished" podID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerID="ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40" exitCode=0 Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.563193 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.563234 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" event={"ID":"a53e86aa-8a32-40af-a371-0055d170c5ce","Type":"ContainerDied","Data":"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.563265 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ddq5p" event={"ID":"a53e86aa-8a32-40af-a371-0055d170c5ce","Type":"ContainerDied","Data":"3f6a9ba8db68876f4db4feda608c526de683aa2e356535c6c54ccdddee549248"} Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.563283 4759 scope.go:117] "RemoveContainer" containerID="ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.565209 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l9lhq" podStartSLOduration=2.567418103 podStartE2EDuration="53.565197546s" podCreationTimestamp="2025-11-25 19:36:08 +0000 UTC" firstStartedPulling="2025-11-25 19:36:09.946859345 +0000 UTC m=+160.532807046" lastFinishedPulling="2025-11-25 19:37:00.944638778 +0000 UTC m=+211.530586489" observedRunningTime="2025-11-25 19:37:01.565052741 +0000 UTC m=+212.151000452" watchObservedRunningTime="2025-11-25 19:37:01.565197546 +0000 UTC m=+212.151145247" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.579907 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.584902 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xxqhr"] Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.585747 4759 scope.go:117] "RemoveContainer" containerID="ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40" Nov 25 19:37:01 crc kubenswrapper[4759]: E1125 19:37:01.586209 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40\": container with ID starting with ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40 not found: ID does not exist" containerID="ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.586236 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40"} err="failed to get container status \"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40\": rpc error: code = NotFound desc = could not find container \"ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40\": container with ID starting with ba29d9b06c0fcc34ee0dd10d5968e2143ff588a222b0d9ea3552530da7307b40 not found: ID does not exist" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.598334 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wfh5l" podStartSLOduration=2.635307862 podStartE2EDuration="51.598319899s" podCreationTimestamp="2025-11-25 19:36:10 +0000 UTC" firstStartedPulling="2025-11-25 19:36:12.001381851 +0000 UTC m=+162.587329562" lastFinishedPulling="2025-11-25 19:37:00.964393898 +0000 UTC m=+211.550341599" observedRunningTime="2025-11-25 19:37:01.596505924 +0000 UTC m=+212.182453625" watchObservedRunningTime="2025-11-25 19:37:01.598319899 +0000 UTC m=+212.184267600" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.631073 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j62pl" podStartSLOduration=2.655198214 podStartE2EDuration="53.631057648s" podCreationTimestamp="2025-11-25 19:36:08 +0000 UTC" firstStartedPulling="2025-11-25 19:36:09.952429386 +0000 UTC m=+160.538377097" lastFinishedPulling="2025-11-25 19:37:00.92828883 +0000 UTC m=+211.514236531" observedRunningTime="2025-11-25 19:37:01.618871646 +0000 UTC m=+212.204819357" watchObservedRunningTime="2025-11-25 19:37:01.631057648 +0000 UTC m=+212.217005349" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.631775 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.634165 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ddq5p"] Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.637691 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.637723 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f9089f5-e196-48ad-9d8c-a14544b6a96f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.637734 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z86ng\" (UniqueName: \"kubernetes.io/projected/6f9089f5-e196-48ad-9d8c-a14544b6a96f-kube-api-access-z86ng\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:01 crc kubenswrapper[4759]: I1125 19:37:01.637744 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9089f5-e196-48ad-9d8c-a14544b6a96f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.112914 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" path="/var/lib/kubelet/pods/6f9089f5-e196-48ad-9d8c-a14544b6a96f/volumes" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.113736 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" path="/var/lib/kubelet/pods/a53e86aa-8a32-40af-a371-0055d170c5ce/volumes" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.813042 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:02 crc kubenswrapper[4759]: E1125 19:37:02.813530 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerName="route-controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.813543 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerName="route-controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: E1125 19:37:02.813556 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerName="controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.813563 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerName="controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.813659 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53e86aa-8a32-40af-a371-0055d170c5ce" containerName="controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.813675 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f9089f5-e196-48ad-9d8c-a14544b6a96f" containerName="route-controller-manager" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.814050 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.821371 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.821816 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.822788 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.822915 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.822934 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.823289 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.824983 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.825652 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830076 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830100 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830231 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830644 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830715 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.830918 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.832091 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.834800 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.837759 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.960860 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw68q\" (UniqueName: \"kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.960929 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.960962 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.960987 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98p2q\" (UniqueName: \"kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.961011 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.961220 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.961257 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.961294 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:02 crc kubenswrapper[4759]: I1125 19:37:02.961330 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062262 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062316 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98p2q\" (UniqueName: \"kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062348 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062888 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062920 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062949 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.062976 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.063036 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw68q\" (UniqueName: \"kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.063066 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.063677 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.063888 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.063953 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.065482 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.066996 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.069100 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.069106 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.078683 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw68q\" (UniqueName: \"kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q\") pod \"controller-manager-6b676fd8cd-ks7cf\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.086795 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98p2q\" (UniqueName: \"kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q\") pod \"route-controller-manager-6d76dfbbdd-xw4gm\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.132635 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.144124 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.347620 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:03 crc kubenswrapper[4759]: W1125 19:37:03.362367 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c8d447_6c87_4c8c_b8cf_9d2915a10105.slice/crio-e33f91d5ae5de40090569765a641805f3b97329610dd6af75ccc8f8778e15a07 WatchSource:0}: Error finding container e33f91d5ae5de40090569765a641805f3b97329610dd6af75ccc8f8778e15a07: Status 404 returned error can't find the container with id e33f91d5ae5de40090569765a641805f3b97329610dd6af75ccc8f8778e15a07 Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.421624 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.577118 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" event={"ID":"cc327e12-b500-4b54-ab71-c1442e564979","Type":"ContainerStarted","Data":"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135"} Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.577159 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" event={"ID":"cc327e12-b500-4b54-ab71-c1442e564979","Type":"ContainerStarted","Data":"a150a4172502bacbdc656c79f70e428f74880e6df76d3b08d9d47b0cd470dd37"} Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.577494 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.579156 4759 patch_prober.go:28] interesting pod/route-controller-manager-6d76dfbbdd-xw4gm container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.579189 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" podUID="cc327e12-b500-4b54-ab71-c1442e564979" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.579533 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" event={"ID":"05c8d447-6c87-4c8c-b8cf-9d2915a10105","Type":"ContainerStarted","Data":"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276"} Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.579559 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" event={"ID":"05c8d447-6c87-4c8c-b8cf-9d2915a10105","Type":"ContainerStarted","Data":"e33f91d5ae5de40090569765a641805f3b97329610dd6af75ccc8f8778e15a07"} Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.579790 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.583211 4759 patch_prober.go:28] interesting pod/controller-manager-6b676fd8cd-ks7cf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.583245 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.604071 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" podStartSLOduration=2.604051432 podStartE2EDuration="2.604051432s" podCreationTimestamp="2025-11-25 19:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:37:03.600209925 +0000 UTC m=+214.186157636" watchObservedRunningTime="2025-11-25 19:37:03.604051432 +0000 UTC m=+214.189999153" Nov 25 19:37:03 crc kubenswrapper[4759]: I1125 19:37:03.619074 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" podStartSLOduration=2.6190543330000002 podStartE2EDuration="2.619054333s" podCreationTimestamp="2025-11-25 19:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:37:03.618864476 +0000 UTC m=+214.204812197" watchObservedRunningTime="2025-11-25 19:37:03.619054333 +0000 UTC m=+214.205002034" Nov 25 19:37:04 crc kubenswrapper[4759]: I1125 19:37:04.588435 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:04 crc kubenswrapper[4759]: I1125 19:37:04.590163 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:08 crc kubenswrapper[4759]: I1125 19:37:08.823129 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:37:08 crc kubenswrapper[4759]: I1125 19:37:08.823676 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:37:08 crc kubenswrapper[4759]: I1125 19:37:08.896082 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.175178 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.215787 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.529639 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.529708 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.580279 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.664531 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:37:09 crc kubenswrapper[4759]: I1125 19:37:09.666586 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:10 crc kubenswrapper[4759]: I1125 19:37:10.526012 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:37:10 crc kubenswrapper[4759]: I1125 19:37:10.825619 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:37:10 crc kubenswrapper[4759]: I1125 19:37:10.825672 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:37:10 crc kubenswrapper[4759]: I1125 19:37:10.862968 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:37:11 crc kubenswrapper[4759]: I1125 19:37:11.523247 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:37:11 crc kubenswrapper[4759]: I1125 19:37:11.523497 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lk6nr" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="registry-server" containerID="cri-o://4f2e40a5b56fcf73d513bc80cc303a26d941ef4e222248207b75fb10c817373d" gracePeriod=2 Nov 25 19:37:11 crc kubenswrapper[4759]: I1125 19:37:11.624364 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j62pl" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="registry-server" containerID="cri-o://3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56" gracePeriod=2 Nov 25 19:37:11 crc kubenswrapper[4759]: I1125 19:37:11.666071 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:37:11 crc kubenswrapper[4759]: E1125 19:37:11.872836 4759 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod967ed015_8b0e_4c14_922f_2df7857bdcb6.slice/crio-3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56.scope\": RecentStats: unable to find data in memory cache]" Nov 25 19:37:12 crc kubenswrapper[4759]: I1125 19:37:12.632547 4759 generic.go:334] "Generic (PLEG): container finished" podID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerID="4f2e40a5b56fcf73d513bc80cc303a26d941ef4e222248207b75fb10c817373d" exitCode=0 Nov 25 19:37:12 crc kubenswrapper[4759]: I1125 19:37:12.632610 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerDied","Data":"4f2e40a5b56fcf73d513bc80cc303a26d941ef4e222248207b75fb10c817373d"} Nov 25 19:37:12 crc kubenswrapper[4759]: I1125 19:37:12.635566 4759 generic.go:334] "Generic (PLEG): container finished" podID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerID="3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56" exitCode=0 Nov 25 19:37:12 crc kubenswrapper[4759]: I1125 19:37:12.635637 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerDied","Data":"3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56"} Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.430600 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.495808 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.505354 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content\") pod \"62dce71f-f0d6-4044-84fd-bc456278eab9\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.505577 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgrjz\" (UniqueName: \"kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz\") pod \"62dce71f-f0d6-4044-84fd-bc456278eab9\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.505657 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities\") pod \"62dce71f-f0d6-4044-84fd-bc456278eab9\" (UID: \"62dce71f-f0d6-4044-84fd-bc456278eab9\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.506319 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities" (OuterVolumeSpecName: "utilities") pod "62dce71f-f0d6-4044-84fd-bc456278eab9" (UID: "62dce71f-f0d6-4044-84fd-bc456278eab9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.515937 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz" (OuterVolumeSpecName: "kube-api-access-fgrjz") pod "62dce71f-f0d6-4044-84fd-bc456278eab9" (UID: "62dce71f-f0d6-4044-84fd-bc456278eab9"). InnerVolumeSpecName "kube-api-access-fgrjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.568121 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62dce71f-f0d6-4044-84fd-bc456278eab9" (UID: "62dce71f-f0d6-4044-84fd-bc456278eab9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607191 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content\") pod \"967ed015-8b0e-4c14-922f-2df7857bdcb6\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607298 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities\") pod \"967ed015-8b0e-4c14-922f-2df7857bdcb6\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607328 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwdhn\" (UniqueName: \"kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn\") pod \"967ed015-8b0e-4c14-922f-2df7857bdcb6\" (UID: \"967ed015-8b0e-4c14-922f-2df7857bdcb6\") " Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607661 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgrjz\" (UniqueName: \"kubernetes.io/projected/62dce71f-f0d6-4044-84fd-bc456278eab9-kube-api-access-fgrjz\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607690 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.607705 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62dce71f-f0d6-4044-84fd-bc456278eab9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.608148 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities" (OuterVolumeSpecName: "utilities") pod "967ed015-8b0e-4c14-922f-2df7857bdcb6" (UID: "967ed015-8b0e-4c14-922f-2df7857bdcb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.611669 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn" (OuterVolumeSpecName: "kube-api-access-qwdhn") pod "967ed015-8b0e-4c14-922f-2df7857bdcb6" (UID: "967ed015-8b0e-4c14-922f-2df7857bdcb6"). InnerVolumeSpecName "kube-api-access-qwdhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.645783 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk6nr" event={"ID":"62dce71f-f0d6-4044-84fd-bc456278eab9","Type":"ContainerDied","Data":"f1b2db4b1690b430e37450c10f0b26ce2d1ba9d89a544c0aa3fe15ac15730699"} Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.645801 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk6nr" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.646663 4759 scope.go:117] "RemoveContainer" containerID="4f2e40a5b56fcf73d513bc80cc303a26d941ef4e222248207b75fb10c817373d" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.649244 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j62pl" event={"ID":"967ed015-8b0e-4c14-922f-2df7857bdcb6","Type":"ContainerDied","Data":"1308f18c4147850715fec36545f3315d47ddab8ffddfd01b0eeae06ddb6fb1f4"} Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.649624 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j62pl" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.656793 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "967ed015-8b0e-4c14-922f-2df7857bdcb6" (UID: "967ed015-8b0e-4c14-922f-2df7857bdcb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.672780 4759 scope.go:117] "RemoveContainer" containerID="992f25b89add659810e7af2e9a548d0caa267305747be7221e0148e19c87b251" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.678553 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.681525 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lk6nr"] Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.703717 4759 scope.go:117] "RemoveContainer" containerID="d53890e421115448a9a299b14ca963d3d3b5c5c014c49fad46ecf81e06a6923d" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.708983 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.709030 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwdhn\" (UniqueName: \"kubernetes.io/projected/967ed015-8b0e-4c14-922f-2df7857bdcb6-kube-api-access-qwdhn\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.709062 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967ed015-8b0e-4c14-922f-2df7857bdcb6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.725645 4759 scope.go:117] "RemoveContainer" containerID="3d9affba69fb8d7871a79ec7c660083e17509f65642e3de71a1e0c8d58073a56" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.742881 4759 scope.go:117] "RemoveContainer" containerID="d7273dbb4cb3477872aa890059f8f5c16ca1d8523150fae7e05b4e4ba3ddbc5a" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.756435 4759 scope.go:117] "RemoveContainer" containerID="613ae1f8df12128c2f4f73baebacb75490d3cac7f47c11366174e3d751873781" Nov 25 19:37:13 crc kubenswrapper[4759]: I1125 19:37:13.998767 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:37:14 crc kubenswrapper[4759]: I1125 19:37:14.001112 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j62pl"] Nov 25 19:37:14 crc kubenswrapper[4759]: I1125 19:37:14.114288 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" path="/var/lib/kubelet/pods/62dce71f-f0d6-4044-84fd-bc456278eab9/volumes" Nov 25 19:37:14 crc kubenswrapper[4759]: I1125 19:37:14.115410 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" path="/var/lib/kubelet/pods/967ed015-8b0e-4c14-922f-2df7857bdcb6/volumes" Nov 25 19:37:20 crc kubenswrapper[4759]: I1125 19:37:20.560609 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pvn5h"] Nov 25 19:37:20 crc kubenswrapper[4759]: I1125 19:37:20.983576 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:20 crc kubenswrapper[4759]: I1125 19:37:20.984233 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerName="controller-manager" containerID="cri-o://506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276" gracePeriod=30 Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.001772 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.001972 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" podUID="cc327e12-b500-4b54-ab71-c1442e564979" containerName="route-controller-manager" containerID="cri-o://de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135" gracePeriod=30 Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.507248 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.571862 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.612398 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca\") pod \"cc327e12-b500-4b54-ab71-c1442e564979\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.612483 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config\") pod \"cc327e12-b500-4b54-ab71-c1442e564979\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.612518 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98p2q\" (UniqueName: \"kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q\") pod \"cc327e12-b500-4b54-ab71-c1442e564979\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.612653 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert\") pod \"cc327e12-b500-4b54-ab71-c1442e564979\" (UID: \"cc327e12-b500-4b54-ab71-c1442e564979\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.613476 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config" (OuterVolumeSpecName: "config") pod "cc327e12-b500-4b54-ab71-c1442e564979" (UID: "cc327e12-b500-4b54-ab71-c1442e564979"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.613540 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca" (OuterVolumeSpecName: "client-ca") pod "cc327e12-b500-4b54-ab71-c1442e564979" (UID: "cc327e12-b500-4b54-ab71-c1442e564979"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.617762 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q" (OuterVolumeSpecName: "kube-api-access-98p2q") pod "cc327e12-b500-4b54-ab71-c1442e564979" (UID: "cc327e12-b500-4b54-ab71-c1442e564979"). InnerVolumeSpecName "kube-api-access-98p2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.617942 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cc327e12-b500-4b54-ab71-c1442e564979" (UID: "cc327e12-b500-4b54-ab71-c1442e564979"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.698276 4759 generic.go:334] "Generic (PLEG): container finished" podID="cc327e12-b500-4b54-ab71-c1442e564979" containerID="de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135" exitCode=0 Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.698335 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" event={"ID":"cc327e12-b500-4b54-ab71-c1442e564979","Type":"ContainerDied","Data":"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135"} Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.698362 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.698390 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm" event={"ID":"cc327e12-b500-4b54-ab71-c1442e564979","Type":"ContainerDied","Data":"a150a4172502bacbdc656c79f70e428f74880e6df76d3b08d9d47b0cd470dd37"} Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.698410 4759 scope.go:117] "RemoveContainer" containerID="de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.699780 4759 generic.go:334] "Generic (PLEG): container finished" podID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerID="506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276" exitCode=0 Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.699812 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" event={"ID":"05c8d447-6c87-4c8c-b8cf-9d2915a10105","Type":"ContainerDied","Data":"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276"} Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.699834 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.699872 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf" event={"ID":"05c8d447-6c87-4c8c-b8cf-9d2915a10105","Type":"ContainerDied","Data":"e33f91d5ae5de40090569765a641805f3b97329610dd6af75ccc8f8778e15a07"} Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713441 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca\") pod \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713539 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles\") pod \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713557 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config\") pod \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713608 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert\") pod \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713642 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw68q\" (UniqueName: \"kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q\") pod \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\" (UID: \"05c8d447-6c87-4c8c-b8cf-9d2915a10105\") " Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713887 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc327e12-b500-4b54-ab71-c1442e564979-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713902 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713910 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc327e12-b500-4b54-ab71-c1442e564979-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.713918 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98p2q\" (UniqueName: \"kubernetes.io/projected/cc327e12-b500-4b54-ab71-c1442e564979-kube-api-access-98p2q\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.714325 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca" (OuterVolumeSpecName: "client-ca") pod "05c8d447-6c87-4c8c-b8cf-9d2915a10105" (UID: "05c8d447-6c87-4c8c-b8cf-9d2915a10105"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.714550 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "05c8d447-6c87-4c8c-b8cf-9d2915a10105" (UID: "05c8d447-6c87-4c8c-b8cf-9d2915a10105"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.714755 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config" (OuterVolumeSpecName: "config") pod "05c8d447-6c87-4c8c-b8cf-9d2915a10105" (UID: "05c8d447-6c87-4c8c-b8cf-9d2915a10105"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.716273 4759 scope.go:117] "RemoveContainer" containerID="de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.716966 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q" (OuterVolumeSpecName: "kube-api-access-fw68q") pod "05c8d447-6c87-4c8c-b8cf-9d2915a10105" (UID: "05c8d447-6c87-4c8c-b8cf-9d2915a10105"). InnerVolumeSpecName "kube-api-access-fw68q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: E1125 19:37:21.719714 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135\": container with ID starting with de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135 not found: ID does not exist" containerID="de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.719769 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135"} err="failed to get container status \"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135\": rpc error: code = NotFound desc = could not find container \"de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135\": container with ID starting with de4114e3c617864e317167ddb4f4d2a9cff1309692a0194d072eb8b041218135 not found: ID does not exist" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.719826 4759 scope.go:117] "RemoveContainer" containerID="506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.720197 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "05c8d447-6c87-4c8c-b8cf-9d2915a10105" (UID: "05c8d447-6c87-4c8c-b8cf-9d2915a10105"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.724332 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.728224 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d76dfbbdd-xw4gm"] Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.735074 4759 scope.go:117] "RemoveContainer" containerID="506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276" Nov 25 19:37:21 crc kubenswrapper[4759]: E1125 19:37:21.735525 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276\": container with ID starting with 506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276 not found: ID does not exist" containerID="506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.735572 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276"} err="failed to get container status \"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276\": rpc error: code = NotFound desc = could not find container \"506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276\": container with ID starting with 506bbeb16fc55c89e184f5311ea643820a846847cefe9260e28a8360db841276 not found: ID does not exist" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.814999 4759 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.815046 4759 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.815067 4759 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c8d447-6c87-4c8c-b8cf-9d2915a10105-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.815412 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw68q\" (UniqueName: \"kubernetes.io/projected/05c8d447-6c87-4c8c-b8cf-9d2915a10105-kube-api-access-fw68q\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:21 crc kubenswrapper[4759]: I1125 19:37:21.815481 4759 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c8d447-6c87-4c8c-b8cf-9d2915a10105-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.071203 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.074333 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b676fd8cd-ks7cf"] Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.113357 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" path="/var/lib/kubelet/pods/05c8d447-6c87-4c8c-b8cf-9d2915a10105/volumes" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.114112 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc327e12-b500-4b54-ab71-c1442e564979" path="/var/lib/kubelet/pods/cc327e12-b500-4b54-ab71-c1442e564979/volumes" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.835194 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6"] Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.835965 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="extract-content" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.835997 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="extract-content" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836024 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836041 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836068 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerName="controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836084 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerName="controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836116 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836131 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836152 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="extract-utilities" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836167 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="extract-utilities" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836185 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc327e12-b500-4b54-ab71-c1442e564979" containerName="route-controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836200 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc327e12-b500-4b54-ab71-c1442e564979" containerName="route-controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836227 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="extract-utilities" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836242 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="extract-utilities" Nov 25 19:37:22 crc kubenswrapper[4759]: E1125 19:37:22.836268 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="extract-content" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836284 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="extract-content" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836520 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8d447-6c87-4c8c-b8cf-9d2915a10105" containerName="controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836564 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="62dce71f-f0d6-4044-84fd-bc456278eab9" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836583 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="967ed015-8b0e-4c14-922f-2df7857bdcb6" containerName="registry-server" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.836602 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc327e12-b500-4b54-ab71-c1442e564979" containerName="route-controller-manager" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.837279 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.840808 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.840840 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.841191 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.841875 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.841893 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.842729 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.844068 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56ddbbc98c-75xq6"] Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.844807 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.847366 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.847876 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.848541 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.848591 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.848551 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.848798 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.855967 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56ddbbc98c-75xq6"] Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.858795 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.870146 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6"] Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.927526 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a75686f7-6265-41f9-b7b9-8390333ecd47-serving-cert\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.927617 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-client-ca\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.927679 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlgb2\" (UniqueName: \"kubernetes.io/projected/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-kube-api-access-qlgb2\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.927759 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-serving-cert\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.927942 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mck4h\" (UniqueName: \"kubernetes.io/projected/a75686f7-6265-41f9-b7b9-8390333ecd47-kube-api-access-mck4h\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.928018 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-client-ca\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.928054 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-proxy-ca-bundles\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.928150 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-config\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:22 crc kubenswrapper[4759]: I1125 19:37:22.928197 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-config\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029325 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-client-ca\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029390 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-proxy-ca-bundles\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029471 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-config\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029516 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-config\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029575 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a75686f7-6265-41f9-b7b9-8390333ecd47-serving-cert\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029618 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-client-ca\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029680 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlgb2\" (UniqueName: \"kubernetes.io/projected/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-kube-api-access-qlgb2\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.029732 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-serving-cert\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.030387 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mck4h\" (UniqueName: \"kubernetes.io/projected/a75686f7-6265-41f9-b7b9-8390333ecd47-kube-api-access-mck4h\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.031686 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-client-ca\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.032173 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-client-ca\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.032410 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a75686f7-6265-41f9-b7b9-8390333ecd47-config\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.033844 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-config\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.042329 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-proxy-ca-bundles\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.042666 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a75686f7-6265-41f9-b7b9-8390333ecd47-serving-cert\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.054118 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlgb2\" (UniqueName: \"kubernetes.io/projected/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-kube-api-access-qlgb2\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.058507 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9af7b02-827e-4f2f-bc8b-aab73e1d6691-serving-cert\") pod \"controller-manager-56ddbbc98c-75xq6\" (UID: \"e9af7b02-827e-4f2f-bc8b-aab73e1d6691\") " pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.062731 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mck4h\" (UniqueName: \"kubernetes.io/projected/a75686f7-6265-41f9-b7b9-8390333ecd47-kube-api-access-mck4h\") pod \"route-controller-manager-7c686c85dc-snsp6\" (UID: \"a75686f7-6265-41f9-b7b9-8390333ecd47\") " pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.166005 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.177885 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.400753 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56ddbbc98c-75xq6"] Nov 25 19:37:23 crc kubenswrapper[4759]: W1125 19:37:23.410431 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9af7b02_827e_4f2f_bc8b_aab73e1d6691.slice/crio-c675e3a059709976b425a15a1bc31f3cdae25cdc884f31b05b72edcb341b5f61 WatchSource:0}: Error finding container c675e3a059709976b425a15a1bc31f3cdae25cdc884f31b05b72edcb341b5f61: Status 404 returned error can't find the container with id c675e3a059709976b425a15a1bc31f3cdae25cdc884f31b05b72edcb341b5f61 Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.677666 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6"] Nov 25 19:37:23 crc kubenswrapper[4759]: W1125 19:37:23.683826 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda75686f7_6265_41f9_b7b9_8390333ecd47.slice/crio-bde4a6357d57fe0727dcde9d1f77f2054459d9e1ca6f0ea81ea5b672f3db5a6e WatchSource:0}: Error finding container bde4a6357d57fe0727dcde9d1f77f2054459d9e1ca6f0ea81ea5b672f3db5a6e: Status 404 returned error can't find the container with id bde4a6357d57fe0727dcde9d1f77f2054459d9e1ca6f0ea81ea5b672f3db5a6e Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.712471 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" event={"ID":"e9af7b02-827e-4f2f-bc8b-aab73e1d6691","Type":"ContainerStarted","Data":"b56bd57a0e91e47f2fab0b5f3c17ffbb35aac02fd2ee532035f60587d0e3264d"} Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.712525 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" event={"ID":"e9af7b02-827e-4f2f-bc8b-aab73e1d6691","Type":"ContainerStarted","Data":"c675e3a059709976b425a15a1bc31f3cdae25cdc884f31b05b72edcb341b5f61"} Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.713034 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.715077 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" event={"ID":"a75686f7-6265-41f9-b7b9-8390333ecd47","Type":"ContainerStarted","Data":"bde4a6357d57fe0727dcde9d1f77f2054459d9e1ca6f0ea81ea5b672f3db5a6e"} Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.718889 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" Nov 25 19:37:23 crc kubenswrapper[4759]: I1125 19:37:23.733628 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56ddbbc98c-75xq6" podStartSLOduration=3.7336036630000002 podStartE2EDuration="3.733603663s" podCreationTimestamp="2025-11-25 19:37:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:37:23.729334712 +0000 UTC m=+234.315282423" watchObservedRunningTime="2025-11-25 19:37:23.733603663 +0000 UTC m=+234.319551374" Nov 25 19:37:24 crc kubenswrapper[4759]: I1125 19:37:24.723861 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" event={"ID":"a75686f7-6265-41f9-b7b9-8390333ecd47","Type":"ContainerStarted","Data":"01de7694740491c0b61e1ffaa8040e14e961793e9262649837c1fb73e331c438"} Nov 25 19:37:24 crc kubenswrapper[4759]: I1125 19:37:24.756032 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" podStartSLOduration=3.756004011 podStartE2EDuration="3.756004011s" podCreationTimestamp="2025-11-25 19:37:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:37:24.748370411 +0000 UTC m=+235.334318122" watchObservedRunningTime="2025-11-25 19:37:24.756004011 +0000 UTC m=+235.341951752" Nov 25 19:37:25 crc kubenswrapper[4759]: I1125 19:37:25.728040 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:25 crc kubenswrapper[4759]: I1125 19:37:25.732900 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c686c85dc-snsp6" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.583892 4759 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.585360 4759 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.585675 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e" gracePeriod=15 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.585886 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.586309 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6" gracePeriod=15 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.586375 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838" gracePeriod=15 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.586416 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9" gracePeriod=15 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.586478 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305" gracePeriod=15 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.589675 4759 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590080 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590115 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590144 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590156 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590173 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590187 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590205 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590219 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590238 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590250 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590272 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590284 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.590304 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590315 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590523 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590554 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590567 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590582 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590600 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.590614 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.635648 4759 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.41:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756409 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756469 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756497 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756521 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756537 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756575 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756636 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.756665 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.795776 4759 generic.go:334] "Generic (PLEG): container finished" podID="254f97f7-20ca-4def-a04d-6ef872a5eb70" containerID="2559372e841f66a750a4a84f8002f6724835e725b9430b9cde0b6fea0d272e8c" exitCode=0 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.795844 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"254f97f7-20ca-4def-a04d-6ef872a5eb70","Type":"ContainerDied","Data":"2559372e841f66a750a4a84f8002f6724835e725b9430b9cde0b6fea0d272e8c"} Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.796401 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.796649 4759 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.798078 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799256 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799852 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6" exitCode=0 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799875 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838" exitCode=0 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799885 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9" exitCode=0 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799895 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305" exitCode=2 Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.799930 4759 scope.go:117] "RemoveContainer" containerID="b912472ba7547d3ec95dcf339508b662e1b3cb94eeee5bcddadb7589f6c212bb" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857511 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857558 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857602 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857617 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857631 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857672 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857680 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857708 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857799 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857715 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857826 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857852 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857876 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857910 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857962 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.857980 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: I1125 19:37:37.936949 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:37 crc kubenswrapper[4759]: W1125 19:37:37.952066 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-ab395a0dfbfdb67362d836e00806c866994bf84dd955bbcc064b40087466fc3d WatchSource:0}: Error finding container ab395a0dfbfdb67362d836e00806c866994bf84dd955bbcc064b40087466fc3d: Status 404 returned error can't find the container with id ab395a0dfbfdb67362d836e00806c866994bf84dd955bbcc064b40087466fc3d Nov 25 19:37:37 crc kubenswrapper[4759]: E1125 19:37:37.954140 4759 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.41:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b571c43ca56f7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 19:37:37.953715959 +0000 UTC m=+248.539663660,LastTimestamp:2025-11-25 19:37:37.953715959 +0000 UTC m=+248.539663660,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 19:37:38 crc kubenswrapper[4759]: I1125 19:37:38.807364 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a"} Nov 25 19:37:38 crc kubenswrapper[4759]: I1125 19:37:38.807415 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ab395a0dfbfdb67362d836e00806c866994bf84dd955bbcc064b40087466fc3d"} Nov 25 19:37:38 crc kubenswrapper[4759]: E1125 19:37:38.808474 4759 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.41:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:38 crc kubenswrapper[4759]: I1125 19:37:38.808499 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:38 crc kubenswrapper[4759]: I1125 19:37:38.811660 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.207182 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.208106 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.376527 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access\") pod \"254f97f7-20ca-4def-a04d-6ef872a5eb70\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.376563 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir\") pod \"254f97f7-20ca-4def-a04d-6ef872a5eb70\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.376598 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock\") pod \"254f97f7-20ca-4def-a04d-6ef872a5eb70\" (UID: \"254f97f7-20ca-4def-a04d-6ef872a5eb70\") " Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.376768 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock" (OuterVolumeSpecName: "var-lock") pod "254f97f7-20ca-4def-a04d-6ef872a5eb70" (UID: "254f97f7-20ca-4def-a04d-6ef872a5eb70"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.377003 4759 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.377060 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "254f97f7-20ca-4def-a04d-6ef872a5eb70" (UID: "254f97f7-20ca-4def-a04d-6ef872a5eb70"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.384818 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "254f97f7-20ca-4def-a04d-6ef872a5eb70" (UID: "254f97f7-20ca-4def-a04d-6ef872a5eb70"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.479140 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/254f97f7-20ca-4def-a04d-6ef872a5eb70-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.479210 4759 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/254f97f7-20ca-4def-a04d-6ef872a5eb70-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.761255 4759 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.762328 4759 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.762697 4759 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.762959 4759 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.763357 4759 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.763574 4759 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.763993 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="200ms" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.818576 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"254f97f7-20ca-4def-a04d-6ef872a5eb70","Type":"ContainerDied","Data":"e70eb6e6a3113750ccda4e9fb3af44719c4b03c38a84d333b3a8fd8721a1bbde"} Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.818621 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e70eb6e6a3113750ccda4e9fb3af44719c4b03c38a84d333b3a8fd8721a1bbde" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.818597 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.819103 4759 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.41:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.830826 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.963924 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 19:37:39 crc kubenswrapper[4759]: E1125 19:37:39.964499 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="400ms" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.964883 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.965266 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:39 crc kubenswrapper[4759]: I1125 19:37:39.965423 4759 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.086644 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.086717 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.086776 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.086931 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.086976 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.087059 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.087386 4759 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.087422 4759 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.087439 4759 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.111211 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.111838 4759 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.114805 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.366313 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="800ms" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.832570 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.834443 4759 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e" exitCode=0 Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.834765 4759 scope.go:117] "RemoveContainer" containerID="de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.835007 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.838082 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.838923 4759 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.840164 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.840973 4759 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.864691 4759 scope.go:117] "RemoveContainer" containerID="a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.882902 4759 scope.go:117] "RemoveContainer" containerID="77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.904186 4759 scope.go:117] "RemoveContainer" containerID="b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.919302 4759 scope.go:117] "RemoveContainer" containerID="338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.934753 4759 scope.go:117] "RemoveContainer" containerID="28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.962485 4759 scope.go:117] "RemoveContainer" containerID="de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.963621 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\": container with ID starting with de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6 not found: ID does not exist" containerID="de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.963663 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6"} err="failed to get container status \"de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\": rpc error: code = NotFound desc = could not find container \"de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6\": container with ID starting with de65b1e91c280133744997e89368ca4fbc0cc3d2afce711fbd8d33e8fd0e31f6 not found: ID does not exist" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.963689 4759 scope.go:117] "RemoveContainer" containerID="a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.964147 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\": container with ID starting with a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838 not found: ID does not exist" containerID="a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964180 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838"} err="failed to get container status \"a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\": rpc error: code = NotFound desc = could not find container \"a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838\": container with ID starting with a45fd259375c741825e330153c1a7dbcd1438f3e1b749cddd3194faf709cc838 not found: ID does not exist" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964202 4759 scope.go:117] "RemoveContainer" containerID="77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.964501 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\": container with ID starting with 77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9 not found: ID does not exist" containerID="77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964524 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9"} err="failed to get container status \"77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\": rpc error: code = NotFound desc = could not find container \"77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9\": container with ID starting with 77279dc156d240b4cb99b67cff264d0ec3efe966e0dbc001149f53e604513af9 not found: ID does not exist" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964540 4759 scope.go:117] "RemoveContainer" containerID="b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.964731 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\": container with ID starting with b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305 not found: ID does not exist" containerID="b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964753 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305"} err="failed to get container status \"b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\": rpc error: code = NotFound desc = could not find container \"b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305\": container with ID starting with b6e224346d1c7d27325ef49507f670569d2e4d2240996b0b44bdf06e92807305 not found: ID does not exist" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964768 4759 scope.go:117] "RemoveContainer" containerID="338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.964966 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\": container with ID starting with 338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e not found: ID does not exist" containerID="338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.964988 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e"} err="failed to get container status \"338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\": rpc error: code = NotFound desc = could not find container \"338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e\": container with ID starting with 338b4a9bf01f05868af40a9402dc86034cdda90c6da8896670e22986f542015e not found: ID does not exist" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.965004 4759 scope.go:117] "RemoveContainer" containerID="28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f" Nov 25 19:37:40 crc kubenswrapper[4759]: E1125 19:37:40.965208 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\": container with ID starting with 28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f not found: ID does not exist" containerID="28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f" Nov 25 19:37:40 crc kubenswrapper[4759]: I1125 19:37:40.965230 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f"} err="failed to get container status \"28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\": rpc error: code = NotFound desc = could not find container \"28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f\": container with ID starting with 28a1c4085e75d8f74f8a6c539e03c1e248bebf97e812a2ec8cb2eb3cd2a9150f not found: ID does not exist" Nov 25 19:37:41 crc kubenswrapper[4759]: E1125 19:37:41.167299 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="1.6s" Nov 25 19:37:42 crc kubenswrapper[4759]: E1125 19:37:42.768571 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="3.2s" Nov 25 19:37:45 crc kubenswrapper[4759]: E1125 19:37:45.499399 4759 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.41:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b571c43ca56f7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 19:37:37.953715959 +0000 UTC m=+248.539663660,LastTimestamp:2025-11-25 19:37:37.953715959 +0000 UTC m=+248.539663660,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 19:37:45 crc kubenswrapper[4759]: I1125 19:37:45.587912 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerName="oauth-openshift" containerID="cri-o://b784f094ad6369d8ce4bdee4e3b06689d0ec68c2e65c2934df19b44b7f10fd5d" gracePeriod=15 Nov 25 19:37:45 crc kubenswrapper[4759]: I1125 19:37:45.866520 4759 generic.go:334] "Generic (PLEG): container finished" podID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerID="b784f094ad6369d8ce4bdee4e3b06689d0ec68c2e65c2934df19b44b7f10fd5d" exitCode=0 Nov 25 19:37:45 crc kubenswrapper[4759]: I1125 19:37:45.866588 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" event={"ID":"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0","Type":"ContainerDied","Data":"b784f094ad6369d8ce4bdee4e3b06689d0ec68c2e65c2934df19b44b7f10fd5d"} Nov 25 19:37:45 crc kubenswrapper[4759]: E1125 19:37:45.970766 4759 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.41:6443: connect: connection refused" interval="6.4s" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.195870 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.196929 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.197631 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.377937 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.378045 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.378681 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.378078 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379183 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379218 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379262 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379285 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379330 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379359 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379398 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379395 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379437 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379591 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379647 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.379730 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62n84\" (UniqueName: \"kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84\") pod \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\" (UID: \"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0\") " Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380666 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380685 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380857 4759 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380883 4759 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380905 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380891 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.380926 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.384677 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.385902 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84" (OuterVolumeSpecName: "kube-api-access-62n84") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "kube-api-access-62n84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.386200 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.386386 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.386681 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.391658 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.391771 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.391969 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.392172 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" (UID: "fcd5d979-97d7-4edc-b64d-b0d3f18c27a0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481693 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62n84\" (UniqueName: \"kubernetes.io/projected/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-kube-api-access-62n84\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481769 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481785 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481798 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481817 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481831 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481845 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481900 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481912 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.481925 4759 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.877039 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" event={"ID":"fcd5d979-97d7-4edc-b64d-b0d3f18c27a0","Type":"ContainerDied","Data":"7bf6a21cc21df0bc844c94891e8c7996a1a3eba6030fba3f3395d3127f25f7db"} Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.877108 4759 scope.go:117] "RemoveContainer" containerID="b784f094ad6369d8ce4bdee4e3b06689d0ec68c2e65c2934df19b44b7f10fd5d" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.877999 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.879030 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.879433 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.896199 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:46 crc kubenswrapper[4759]: I1125 19:37:46.896772 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.106789 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.110730 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.111159 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.112001 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.112292 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.134503 4759 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.134541 4759 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:50 crc kubenswrapper[4759]: E1125 19:37:50.135052 4759 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.135363 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.906852 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.906942 4759 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e" exitCode=1 Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.907041 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e"} Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.907825 4759 scope.go:117] "RemoveContainer" containerID="ff4a976a354a7a082fa5f6ca6d538dec89a4c0e5b46eb6067dbc28be1497766e" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908161 4759 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908595 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908858 4759 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="4d512e6ea2ccb27082d1a62c7a90e4a49164cb7a7563449e0d282d029cbb8007" exitCode=0 Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908892 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"4d512e6ea2ccb27082d1a62c7a90e4a49164cb7a7563449e0d282d029cbb8007"} Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908912 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dfb97ae5feb5713a684edd36286d5fe14f1cbe8a63c7f4abd54e82629ed1a889"} Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.908983 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.909143 4759 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.909163 4759 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.909658 4759 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: E1125 19:37:50.909727 4759 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.910079 4759 status_manager.go:851] "Failed to get status for pod" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:50 crc kubenswrapper[4759]: I1125 19:37:50.910757 4759 status_manager.go:851] "Failed to get status for pod" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" pod="openshift-authentication/oauth-openshift-558db77b4-pvn5h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pvn5h\": dial tcp 38.102.83.41:6443: connect: connection refused" Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.919248 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bd5dcfbd6532bc70e45d626d30ce61fb816a75c4b9a928ab443801eba74889e2"} Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.919529 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0b2f232e0b10bdde423b8bd989248c868de2a0eff0d11696b3480e006634955f"} Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.919539 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e7db32d475e6ff5a6b7ae98feb0743c2eaccde14502d196177eceee7347a4f4a"} Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.919548 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cd649e5316377f8650a5b4b8751a472961f1b832d91fdf5f2bd6626edb30c7f2"} Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.922927 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 19:37:51 crc kubenswrapper[4759]: I1125 19:37:51.923207 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"690f10664174237bd7a87582274f11b924f5854d439ca709deafd1386934208a"} Nov 25 19:37:52 crc kubenswrapper[4759]: I1125 19:37:52.931098 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"517a29c067139ec12dbdf2f472cc7cf24d71ca6ec489938e7abefa6056431ef5"} Nov 25 19:37:52 crc kubenswrapper[4759]: I1125 19:37:52.931589 4759 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:52 crc kubenswrapper[4759]: I1125 19:37:52.931607 4759 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:52 crc kubenswrapper[4759]: I1125 19:37:52.931908 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.135983 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.136320 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.141917 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.559528 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.569991 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:37:55 crc kubenswrapper[4759]: I1125 19:37:55.817866 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:37:57 crc kubenswrapper[4759]: I1125 19:37:57.945364 4759 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:57 crc kubenswrapper[4759]: I1125 19:37:57.964694 4759 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:57 crc kubenswrapper[4759]: I1125 19:37:57.964723 4759 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:57 crc kubenswrapper[4759]: I1125 19:37:57.969430 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:37:58 crc kubenswrapper[4759]: I1125 19:37:58.970361 4759 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:37:58 crc kubenswrapper[4759]: I1125 19:37:58.970391 4759 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b9b1f292-6844-43a6-bb02-a0e661e3061e" Nov 25 19:38:00 crc kubenswrapper[4759]: I1125 19:38:00.128155 4759 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f6789e5a-96a0-4af1-9205-aec941077093" Nov 25 19:38:05 crc kubenswrapper[4759]: I1125 19:38:05.824570 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 19:38:08 crc kubenswrapper[4759]: I1125 19:38:08.366252 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 19:38:08 crc kubenswrapper[4759]: I1125 19:38:08.549785 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 19:38:08 crc kubenswrapper[4759]: I1125 19:38:08.631192 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 19:38:08 crc kubenswrapper[4759]: I1125 19:38:08.752797 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 19:38:09 crc kubenswrapper[4759]: I1125 19:38:09.070169 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 19:38:09 crc kubenswrapper[4759]: I1125 19:38:09.270019 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 19:38:09 crc kubenswrapper[4759]: I1125 19:38:09.617036 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 19:38:09 crc kubenswrapper[4759]: I1125 19:38:09.648214 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 19:38:09 crc kubenswrapper[4759]: I1125 19:38:09.838258 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.241077 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.415339 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.503536 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.637584 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.763047 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 19:38:10 crc kubenswrapper[4759]: I1125 19:38:10.852571 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.125175 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.228482 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.280531 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.410617 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.490691 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.508757 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.623826 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.628740 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.785348 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 19:38:11 crc kubenswrapper[4759]: I1125 19:38:11.908261 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.098041 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.135845 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.239936 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.355950 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.455808 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.457207 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.494517 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.499294 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.544666 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.561413 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.624664 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.726399 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.827626 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.914420 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 19:38:12 crc kubenswrapper[4759]: I1125 19:38:12.996793 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.353748 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.389596 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.393627 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.427282 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.506747 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.514927 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.603085 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.740813 4759 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.742116 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.745838 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-pvn5h"] Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.745903 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.751563 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.769385 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.769370941 podStartE2EDuration="16.769370941s" podCreationTimestamp="2025-11-25 19:37:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:38:13.766609644 +0000 UTC m=+284.352557375" watchObservedRunningTime="2025-11-25 19:38:13.769370941 +0000 UTC m=+284.355318642" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.860586 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.899651 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.917838 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 19:38:13 crc kubenswrapper[4759]: I1125 19:38:13.931809 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.001683 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.026364 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.042086 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.113515 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" path="/var/lib/kubelet/pods/fcd5d979-97d7-4edc-b64d-b0d3f18c27a0/volumes" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.148969 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.187501 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.195691 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.263657 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.290853 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf"] Nov 25 19:38:14 crc kubenswrapper[4759]: E1125 19:38:14.291158 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" containerName="installer" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.291185 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" containerName="installer" Nov 25 19:38:14 crc kubenswrapper[4759]: E1125 19:38:14.291207 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerName="oauth-openshift" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.291220 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerName="oauth-openshift" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.291403 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd5d979-97d7-4edc-b64d-b0d3f18c27a0" containerName="oauth-openshift" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.291429 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="254f97f7-20ca-4def-a04d-6ef872a5eb70" containerName="installer" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.292061 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.294206 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.294255 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.294409 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.294507 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.294690 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.295154 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.295331 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.296302 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.296630 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.296739 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.296656 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.297772 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.303075 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.312516 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.316846 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.423994 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424257 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424358 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-dir\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424464 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424554 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-policies\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424645 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424720 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424808 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424909 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.424994 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.425079 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbv5q\" (UniqueName: \"kubernetes.io/projected/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-kube-api-access-fbv5q\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.425159 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.425229 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.425298 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.427401 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.431607 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.526563 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.526650 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527609 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbv5q\" (UniqueName: \"kubernetes.io/projected/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-kube-api-access-fbv5q\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527699 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527733 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527768 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527824 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527859 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527904 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-dir\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527935 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527959 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-policies\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.527985 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.528007 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.528034 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.528819 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-dir\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.529176 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.529330 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.529557 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.529777 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-audit-policies\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.533539 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.534344 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.534413 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.534715 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.535244 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.536784 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.537713 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.537990 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.544858 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbv5q\" (UniqueName: \"kubernetes.io/projected/2e9deba3-4b35-4554-b470-9aa9e40dcdbe-kube-api-access-fbv5q\") pod \"oauth-openshift-5f7dff4c75-j5gjf\" (UID: \"2e9deba3-4b35-4554-b470-9aa9e40dcdbe\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.584990 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.613605 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:14 crc kubenswrapper[4759]: I1125 19:38:14.806735 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.061982 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.248905 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.290951 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.381190 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.409079 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.459971 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.521814 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.607088 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.694238 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.709697 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.730753 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.739555 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.811849 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.930253 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.950283 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 19:38:15 crc kubenswrapper[4759]: I1125 19:38:15.966872 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.100323 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.115065 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.207768 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.211336 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.253804 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.265656 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.281509 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.335455 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.339895 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.352020 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.381400 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.647892 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.836573 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.844140 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.927925 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 19:38:16 crc kubenswrapper[4759]: I1125 19:38:16.959497 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.043751 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.082754 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.098737 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.099161 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.124215 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.126430 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.153096 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.180857 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.254294 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.318878 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.331487 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.349871 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.356747 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.377958 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.475816 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.558236 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.577381 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.579130 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.589419 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.602752 4759 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.625121 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.645519 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.655642 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.679437 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.754581 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.755381 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.849773 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 19:38:17 crc kubenswrapper[4759]: I1125 19:38:17.948298 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.020835 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.082023 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.082661 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.159529 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.181340 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.205176 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.207336 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.210095 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.217589 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.221797 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.310230 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.331325 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.409119 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.449905 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.632163 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.644656 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.672943 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.708991 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.931116 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 19:38:18 crc kubenswrapper[4759]: I1125 19:38:18.942254 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.017590 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.174373 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.178330 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.190323 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.205452 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.234755 4759 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.274169 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.449355 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.527079 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.563046 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.577933 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.648581 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.663321 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.697899 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.777928 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.797619 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.803042 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.927530 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.927698 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 19:38:19 crc kubenswrapper[4759]: I1125 19:38:19.936925 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.040667 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.064514 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.084649 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.145579 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.163164 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.189716 4759 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.190972 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a" gracePeriod=5 Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.397181 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.479048 4759 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.581626 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.586775 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.596125 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.700121 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.740311 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.773171 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.812334 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.866956 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.879929 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.914308 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.924750 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.988253 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 19:38:20 crc kubenswrapper[4759]: I1125 19:38:20.991823 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.074306 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.091844 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.098158 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.232245 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.318193 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.328706 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.381533 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.469487 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.518648 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.645410 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.705002 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.771710 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.824165 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.884605 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 19:38:21 crc kubenswrapper[4759]: I1125 19:38:21.964186 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.041570 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.055062 4759 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.179019 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.211160 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.212730 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.281791 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.374568 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.414528 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.461375 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.481545 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.529510 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.553099 4759 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.594572 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.624851 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.625736 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.693287 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.696497 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.843134 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf"] Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.940490 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 19:38:22 crc kubenswrapper[4759]: I1125 19:38:22.984719 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.069217 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.276030 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.278866 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf"] Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.358603 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.459399 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.536101 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.594055 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 19:38:23 crc kubenswrapper[4759]: I1125 19:38:23.734928 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.071129 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.094872 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.114532 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" event={"ID":"2e9deba3-4b35-4554-b470-9aa9e40dcdbe","Type":"ContainerStarted","Data":"ddcd51daebf6d1a021f12e7767c190cf74162e635feedbea37df35a9d89fbf9e"} Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.114592 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.114612 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" event={"ID":"2e9deba3-4b35-4554-b470-9aa9e40dcdbe","Type":"ContainerStarted","Data":"71efb4ddc86accc75ff4d5906ad9c0d40fb3071a3e45ac45a9683d26e603b560"} Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.116840 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.169848 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.190965 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.207138 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5f7dff4c75-j5gjf" podStartSLOduration=64.207121391 podStartE2EDuration="1m4.207121391s" podCreationTimestamp="2025-11-25 19:37:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:38:24.135185181 +0000 UTC m=+294.721132922" watchObservedRunningTime="2025-11-25 19:38:24.207121391 +0000 UTC m=+294.793069092" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.321272 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.338507 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.484648 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 19:38:24 crc kubenswrapper[4759]: I1125 19:38:24.495135 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.361913 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.751360 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.751429 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.838523 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895018 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895383 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895434 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895506 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895575 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895632 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895727 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895752 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895808 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.895888 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.896188 4759 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.896205 4759 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.896214 4759 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.896224 4759 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.903028 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:38:25 crc kubenswrapper[4759]: I1125 19:38:25.998470 4759 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.093895 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.149899 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.160301 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.160368 4759 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a" exitCode=137 Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.161405 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.162580 4759 scope.go:117] "RemoveContainer" containerID="d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.189892 4759 scope.go:117] "RemoveContainer" containerID="d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a" Nov 25 19:38:26 crc kubenswrapper[4759]: E1125 19:38:26.190796 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a\": container with ID starting with d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a not found: ID does not exist" containerID="d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a" Nov 25 19:38:26 crc kubenswrapper[4759]: I1125 19:38:26.190872 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a"} err="failed to get container status \"d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a\": rpc error: code = NotFound desc = could not find container \"d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a\": container with ID starting with d93b2053c996436890561f4c8879ef1cc3abf4a377e50d69fb3260690289174a not found: ID does not exist" Nov 25 19:39:24 crc kubenswrapper[4759]: I1125 19:39:24.895726 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:39:24 crc kubenswrapper[4759]: I1125 19:39:24.896319 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.463163 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fdbg2"] Nov 25 19:39:50 crc kubenswrapper[4759]: E1125 19:39:50.463966 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.463982 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.464115 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.464772 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.473396 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fdbg2"] Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576005 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-tls\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576063 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kzgz\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-kube-api-access-4kzgz\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576083 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86103804-a2e1-40cb-b9e1-9f852aa32c83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576101 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-trusted-ca\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576162 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576332 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86103804-a2e1-40cb-b9e1-9f852aa32c83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576382 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-bound-sa-token\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.576405 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-certificates\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.594114 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.677922 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86103804-a2e1-40cb-b9e1-9f852aa32c83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.677965 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-bound-sa-token\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.677986 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-certificates\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.678016 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-tls\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.678044 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kzgz\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-kube-api-access-4kzgz\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.678065 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86103804-a2e1-40cb-b9e1-9f852aa32c83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.678087 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-trusted-ca\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.678920 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86103804-a2e1-40cb-b9e1-9f852aa32c83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.679139 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-certificates\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.679533 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86103804-a2e1-40cb-b9e1-9f852aa32c83-trusted-ca\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.683536 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86103804-a2e1-40cb-b9e1-9f852aa32c83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.683579 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-registry-tls\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.695084 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kzgz\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-kube-api-access-4kzgz\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.696070 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86103804-a2e1-40cb-b9e1-9f852aa32c83-bound-sa-token\") pod \"image-registry-66df7c8f76-fdbg2\" (UID: \"86103804-a2e1-40cb-b9e1-9f852aa32c83\") " pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.786702 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:50 crc kubenswrapper[4759]: I1125 19:39:50.967842 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fdbg2"] Nov 25 19:39:51 crc kubenswrapper[4759]: I1125 19:39:51.650000 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" event={"ID":"86103804-a2e1-40cb-b9e1-9f852aa32c83","Type":"ContainerStarted","Data":"3df6fb8b466db1eebf04d0af9e46528d231ffd8eff415124b8627b8b769cb50a"} Nov 25 19:39:51 crc kubenswrapper[4759]: I1125 19:39:51.650313 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" event={"ID":"86103804-a2e1-40cb-b9e1-9f852aa32c83","Type":"ContainerStarted","Data":"db922ac95a844c6339d80c087b7b1dca9f77e8b89753751028fd435c0b497bef"} Nov 25 19:39:51 crc kubenswrapper[4759]: I1125 19:39:51.650329 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:39:51 crc kubenswrapper[4759]: I1125 19:39:51.677645 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" podStartSLOduration=1.677613391 podStartE2EDuration="1.677613391s" podCreationTimestamp="2025-11-25 19:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:39:51.665573561 +0000 UTC m=+382.251521342" watchObservedRunningTime="2025-11-25 19:39:51.677613391 +0000 UTC m=+382.263561132" Nov 25 19:39:54 crc kubenswrapper[4759]: I1125 19:39:54.895842 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:39:54 crc kubenswrapper[4759]: I1125 19:39:54.896141 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:40:10 crc kubenswrapper[4759]: I1125 19:40:10.793627 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fdbg2" Nov 25 19:40:10 crc kubenswrapper[4759]: I1125 19:40:10.859943 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.812721 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.813434 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l9lhq" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="registry-server" containerID="cri-o://166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1" gracePeriod=30 Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.816378 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.816699 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nz425" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="registry-server" containerID="cri-o://b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983" gracePeriod=30 Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.827072 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.827286 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" containerID="cri-o://72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7" gracePeriod=30 Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.833530 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.834021 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wfh5l" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="registry-server" containerID="cri-o://a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd" gracePeriod=30 Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.837912 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.838183 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6zx9x" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="registry-server" containerID="cri-o://cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86" gracePeriod=30 Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.853458 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tvgvw"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.869537 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.887224 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tvgvw"] Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.912843 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.912928 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:19 crc kubenswrapper[4759]: I1125 19:40:19.912990 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gql8w\" (UniqueName: \"kubernetes.io/projected/95673284-6f81-4c46-8fa6-825dc24d77b1-kube-api-access-gql8w\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.014362 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.014465 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.014500 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gql8w\" (UniqueName: \"kubernetes.io/projected/95673284-6f81-4c46-8fa6-825dc24d77b1-kube-api-access-gql8w\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.019172 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.023055 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95673284-6f81-4c46-8fa6-825dc24d77b1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.041272 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gql8w\" (UniqueName: \"kubernetes.io/projected/95673284-6f81-4c46-8fa6-825dc24d77b1-kube-api-access-gql8w\") pod \"marketplace-operator-79b997595-tvgvw\" (UID: \"95673284-6f81-4c46-8fa6-825dc24d77b1\") " pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.219857 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.224828 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.233110 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.259229 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.272990 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.299311 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317669 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca\") pod \"ac427820-58ca-4e4f-b8c6-058084c79617\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317715 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities\") pod \"401e8bc0-19b1-440e-8062-298d61c12e37\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317742 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content\") pod \"a685e63a-c301-4410-84f7-6825651ad294\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317791 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxkst\" (UniqueName: \"kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst\") pod \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317832 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl5p2\" (UniqueName: \"kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2\") pod \"3eb78bd8-c3b4-4daa-a818-07501af8098f\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317849 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content\") pod \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317873 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcsg8\" (UniqueName: \"kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8\") pod \"a685e63a-c301-4410-84f7-6825651ad294\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317892 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities\") pod \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\" (UID: \"d0ded011-6d23-4501-a7ea-f90ed6696cdd\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317906 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities\") pod \"a685e63a-c301-4410-84f7-6825651ad294\" (UID: \"a685e63a-c301-4410-84f7-6825651ad294\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317924 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content\") pod \"401e8bc0-19b1-440e-8062-298d61c12e37\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317951 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf6rn\" (UniqueName: \"kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn\") pod \"401e8bc0-19b1-440e-8062-298d61c12e37\" (UID: \"401e8bc0-19b1-440e-8062-298d61c12e37\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.317991 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics\") pod \"ac427820-58ca-4e4f-b8c6-058084c79617\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.318013 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities\") pod \"3eb78bd8-c3b4-4daa-a818-07501af8098f\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.318040 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cht75\" (UniqueName: \"kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75\") pod \"ac427820-58ca-4e4f-b8c6-058084c79617\" (UID: \"ac427820-58ca-4e4f-b8c6-058084c79617\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.318059 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content\") pod \"3eb78bd8-c3b4-4daa-a818-07501af8098f\" (UID: \"3eb78bd8-c3b4-4daa-a818-07501af8098f\") " Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.319952 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities" (OuterVolumeSpecName: "utilities") pod "3eb78bd8-c3b4-4daa-a818-07501af8098f" (UID: "3eb78bd8-c3b4-4daa-a818-07501af8098f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.320851 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities" (OuterVolumeSpecName: "utilities") pod "a685e63a-c301-4410-84f7-6825651ad294" (UID: "a685e63a-c301-4410-84f7-6825651ad294"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.322232 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ac427820-58ca-4e4f-b8c6-058084c79617" (UID: "ac427820-58ca-4e4f-b8c6-058084c79617"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.323314 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ac427820-58ca-4e4f-b8c6-058084c79617" (UID: "ac427820-58ca-4e4f-b8c6-058084c79617"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.323314 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst" (OuterVolumeSpecName: "kube-api-access-gxkst") pod "d0ded011-6d23-4501-a7ea-f90ed6696cdd" (UID: "d0ded011-6d23-4501-a7ea-f90ed6696cdd"). InnerVolumeSpecName "kube-api-access-gxkst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.325215 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities" (OuterVolumeSpecName: "utilities") pod "401e8bc0-19b1-440e-8062-298d61c12e37" (UID: "401e8bc0-19b1-440e-8062-298d61c12e37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.325226 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75" (OuterVolumeSpecName: "kube-api-access-cht75") pod "ac427820-58ca-4e4f-b8c6-058084c79617" (UID: "ac427820-58ca-4e4f-b8c6-058084c79617"). InnerVolumeSpecName "kube-api-access-cht75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.332554 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities" (OuterVolumeSpecName: "utilities") pod "d0ded011-6d23-4501-a7ea-f90ed6696cdd" (UID: "d0ded011-6d23-4501-a7ea-f90ed6696cdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.332565 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn" (OuterVolumeSpecName: "kube-api-access-cf6rn") pod "401e8bc0-19b1-440e-8062-298d61c12e37" (UID: "401e8bc0-19b1-440e-8062-298d61c12e37"). InnerVolumeSpecName "kube-api-access-cf6rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.335266 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8" (OuterVolumeSpecName: "kube-api-access-fcsg8") pod "a685e63a-c301-4410-84f7-6825651ad294" (UID: "a685e63a-c301-4410-84f7-6825651ad294"). InnerVolumeSpecName "kube-api-access-fcsg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.335599 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2" (OuterVolumeSpecName: "kube-api-access-wl5p2") pod "3eb78bd8-c3b4-4daa-a818-07501af8098f" (UID: "3eb78bd8-c3b4-4daa-a818-07501af8098f"). InnerVolumeSpecName "kube-api-access-wl5p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.365157 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a685e63a-c301-4410-84f7-6825651ad294" (UID: "a685e63a-c301-4410-84f7-6825651ad294"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.382939 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "401e8bc0-19b1-440e-8062-298d61c12e37" (UID: "401e8bc0-19b1-440e-8062-298d61c12e37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.390734 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0ded011-6d23-4501-a7ea-f90ed6696cdd" (UID: "d0ded011-6d23-4501-a7ea-f90ed6696cdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.401672 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tvgvw"] Nov 25 19:40:20 crc kubenswrapper[4759]: W1125 19:40:20.411211 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95673284_6f81_4c46_8fa6_825dc24d77b1.slice/crio-e369b50027a6a1a19f9bf21cd31a55d59cb67c49a519d734dba84dc9b13a19e7 WatchSource:0}: Error finding container e369b50027a6a1a19f9bf21cd31a55d59cb67c49a519d734dba84dc9b13a19e7: Status 404 returned error can't find the container with id e369b50027a6a1a19f9bf21cd31a55d59cb67c49a519d734dba84dc9b13a19e7 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421074 4759 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421462 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421489 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cht75\" (UniqueName: \"kubernetes.io/projected/ac427820-58ca-4e4f-b8c6-058084c79617-kube-api-access-cht75\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421503 4759 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac427820-58ca-4e4f-b8c6-058084c79617-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421518 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421531 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421545 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxkst\" (UniqueName: \"kubernetes.io/projected/d0ded011-6d23-4501-a7ea-f90ed6696cdd-kube-api-access-gxkst\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421558 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421572 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl5p2\" (UniqueName: \"kubernetes.io/projected/3eb78bd8-c3b4-4daa-a818-07501af8098f-kube-api-access-wl5p2\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421586 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcsg8\" (UniqueName: \"kubernetes.io/projected/a685e63a-c301-4410-84f7-6825651ad294-kube-api-access-fcsg8\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421598 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0ded011-6d23-4501-a7ea-f90ed6696cdd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421611 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a685e63a-c301-4410-84f7-6825651ad294-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421623 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401e8bc0-19b1-440e-8062-298d61c12e37-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.421635 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf6rn\" (UniqueName: \"kubernetes.io/projected/401e8bc0-19b1-440e-8062-298d61c12e37-kube-api-access-cf6rn\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.458324 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3eb78bd8-c3b4-4daa-a818-07501af8098f" (UID: "3eb78bd8-c3b4-4daa-a818-07501af8098f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.522509 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eb78bd8-c3b4-4daa-a818-07501af8098f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.818471 4759 generic.go:334] "Generic (PLEG): container finished" podID="401e8bc0-19b1-440e-8062-298d61c12e37" containerID="b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983" exitCode=0 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.818563 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz425" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.818553 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerDied","Data":"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.818674 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz425" event={"ID":"401e8bc0-19b1-440e-8062-298d61c12e37","Type":"ContainerDied","Data":"7c17d178803fcaa7a9838f1be9d8879f82b6aefa7464dd5007ad4a5cb4b80653"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.818695 4759 scope.go:117] "RemoveContainer" containerID="b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.821597 4759 generic.go:334] "Generic (PLEG): container finished" podID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerID="166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1" exitCode=0 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.821655 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9lhq" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.821682 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerDied","Data":"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.821717 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9lhq" event={"ID":"d0ded011-6d23-4501-a7ea-f90ed6696cdd","Type":"ContainerDied","Data":"87c5cd3110ba1e97a2a7a37c24e68286e95e9727aa1d4a18e2e0a82f5848cc16"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.825251 4759 generic.go:334] "Generic (PLEG): container finished" podID="a685e63a-c301-4410-84f7-6825651ad294" containerID="a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd" exitCode=0 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.825330 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerDied","Data":"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.825358 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfh5l" event={"ID":"a685e63a-c301-4410-84f7-6825651ad294","Type":"ContainerDied","Data":"c1b7c182aaffc478ae9f103c9856fc7ee741bd0ce27403034ad4c98a46df3930"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.825367 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfh5l" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.826873 4759 generic.go:334] "Generic (PLEG): container finished" podID="ac427820-58ca-4e4f-b8c6-058084c79617" containerID="72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7" exitCode=0 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.826935 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" event={"ID":"ac427820-58ca-4e4f-b8c6-058084c79617","Type":"ContainerDied","Data":"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.826958 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" event={"ID":"ac427820-58ca-4e4f-b8c6-058084c79617","Type":"ContainerDied","Data":"271e0baa3eea51ecb3de3590c4a7d2072d0159b805e814457fd7de870df4a7ba"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.827009 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwszg" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.830899 4759 generic.go:334] "Generic (PLEG): container finished" podID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerID="cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86" exitCode=0 Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.830969 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerDied","Data":"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.830985 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6zx9x" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.831004 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6zx9x" event={"ID":"3eb78bd8-c3b4-4daa-a818-07501af8098f","Type":"ContainerDied","Data":"096f077670c195bb8b992069ffe3871b61d623e927a8f263abdbd3ed66f2c59b"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.832673 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" event={"ID":"95673284-6f81-4c46-8fa6-825dc24d77b1","Type":"ContainerStarted","Data":"0e3372c5a0479206a385e2d56b13a25e65e86fc6403560ba669382640b60f66d"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.832732 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" event={"ID":"95673284-6f81-4c46-8fa6-825dc24d77b1","Type":"ContainerStarted","Data":"e369b50027a6a1a19f9bf21cd31a55d59cb67c49a519d734dba84dc9b13a19e7"} Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.833660 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.834192 4759 scope.go:117] "RemoveContainer" containerID="2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.837477 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.860965 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tvgvw" podStartSLOduration=1.86094428 podStartE2EDuration="1.86094428s" podCreationTimestamp="2025-11-25 19:40:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:40:20.853622108 +0000 UTC m=+411.439569829" watchObservedRunningTime="2025-11-25 19:40:20.86094428 +0000 UTC m=+411.446892001" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.870797 4759 scope.go:117] "RemoveContainer" containerID="1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.871501 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.875664 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nz425"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.905843 4759 scope.go:117] "RemoveContainer" containerID="b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.906396 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.910112 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l9lhq"] Nov 25 19:40:20 crc kubenswrapper[4759]: E1125 19:40:20.916620 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983\": container with ID starting with b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983 not found: ID does not exist" containerID="b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.916680 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983"} err="failed to get container status \"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983\": rpc error: code = NotFound desc = could not find container \"b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983\": container with ID starting with b8da15564b89cf78e56dbdb93aae8bcb132fd3c5683cf0cea8ef2d38f0840983 not found: ID does not exist" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.916714 4759 scope.go:117] "RemoveContainer" containerID="2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6" Nov 25 19:40:20 crc kubenswrapper[4759]: E1125 19:40:20.918723 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6\": container with ID starting with 2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6 not found: ID does not exist" containerID="2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.918766 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6"} err="failed to get container status \"2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6\": rpc error: code = NotFound desc = could not find container \"2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6\": container with ID starting with 2c993a7f2cf6b19e05c44185e990aaf98ba21b6cd38788b1ffd851be6566c0a6 not found: ID does not exist" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.918797 4759 scope.go:117] "RemoveContainer" containerID="1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069" Nov 25 19:40:20 crc kubenswrapper[4759]: E1125 19:40:20.919166 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069\": container with ID starting with 1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069 not found: ID does not exist" containerID="1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.919228 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069"} err="failed to get container status \"1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069\": rpc error: code = NotFound desc = could not find container \"1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069\": container with ID starting with 1c2a17eab89eeffc2c1586e186f8772145cad6da64e91bfe1cc2af4a11ec1069 not found: ID does not exist" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.919279 4759 scope.go:117] "RemoveContainer" containerID="166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1" Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.930394 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.944509 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwszg"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.948867 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.954253 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6zx9x"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.959422 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.975928 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfh5l"] Nov 25 19:40:20 crc kubenswrapper[4759]: I1125 19:40:20.990033 4759 scope.go:117] "RemoveContainer" containerID="598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.020889 4759 scope.go:117] "RemoveContainer" containerID="bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.038595 4759 scope.go:117] "RemoveContainer" containerID="166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.039101 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1\": container with ID starting with 166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1 not found: ID does not exist" containerID="166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.039145 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1"} err="failed to get container status \"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1\": rpc error: code = NotFound desc = could not find container \"166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1\": container with ID starting with 166007c4c3d13bbb7a68d5d136fea96f9565fece1e491509a9613592124230e1 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.039174 4759 scope.go:117] "RemoveContainer" containerID="598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.039526 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073\": container with ID starting with 598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073 not found: ID does not exist" containerID="598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.039570 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073"} err="failed to get container status \"598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073\": rpc error: code = NotFound desc = could not find container \"598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073\": container with ID starting with 598c8b5d1d6925042003efc8c83afd84d200b31d71cf2ce8646d1b28d3a23073 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.039600 4759 scope.go:117] "RemoveContainer" containerID="bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.039988 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8\": container with ID starting with bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8 not found: ID does not exist" containerID="bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.040019 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8"} err="failed to get container status \"bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8\": rpc error: code = NotFound desc = could not find container \"bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8\": container with ID starting with bcdbe5937c652992e4e1d5a58ebb49fb09771f4ee0448f5df47a098c587294a8 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.040038 4759 scope.go:117] "RemoveContainer" containerID="a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.052140 4759 scope.go:117] "RemoveContainer" containerID="3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.064788 4759 scope.go:117] "RemoveContainer" containerID="d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.081513 4759 scope.go:117] "RemoveContainer" containerID="a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.082116 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd\": container with ID starting with a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd not found: ID does not exist" containerID="a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.082160 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd"} err="failed to get container status \"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd\": rpc error: code = NotFound desc = could not find container \"a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd\": container with ID starting with a9e68ab549f8d0e726ff63402d2846f4ad37715e0a735ac3003780cbedc8dddd not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.082187 4759 scope.go:117] "RemoveContainer" containerID="3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.082743 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3\": container with ID starting with 3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3 not found: ID does not exist" containerID="3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.082793 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3"} err="failed to get container status \"3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3\": rpc error: code = NotFound desc = could not find container \"3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3\": container with ID starting with 3a389dec5ce794aa58c4ad50f4c8ace74274667334f4bed7b25f0b5821c568b3 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.082829 4759 scope.go:117] "RemoveContainer" containerID="d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.083120 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c\": container with ID starting with d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c not found: ID does not exist" containerID="d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.083143 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c"} err="failed to get container status \"d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c\": rpc error: code = NotFound desc = could not find container \"d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c\": container with ID starting with d73ed632a0e22e73812807f40234e824294f90d8f4b3bf21eb0c066b1278f53c not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.083157 4759 scope.go:117] "RemoveContainer" containerID="72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.095293 4759 scope.go:117] "RemoveContainer" containerID="72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.095615 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7\": container with ID starting with 72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7 not found: ID does not exist" containerID="72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.095657 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7"} err="failed to get container status \"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7\": rpc error: code = NotFound desc = could not find container \"72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7\": container with ID starting with 72276d1e6d2e7273d2f44fd2474d8d5bc39a5c44f903b92b50385e10d598f3a7 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.095703 4759 scope.go:117] "RemoveContainer" containerID="cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.107324 4759 scope.go:117] "RemoveContainer" containerID="fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.121564 4759 scope.go:117] "RemoveContainer" containerID="9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.133684 4759 scope.go:117] "RemoveContainer" containerID="cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.134037 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86\": container with ID starting with cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86 not found: ID does not exist" containerID="cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.134081 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86"} err="failed to get container status \"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86\": rpc error: code = NotFound desc = could not find container \"cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86\": container with ID starting with cd2b6ca3977eb41f67d54979e1cb6cc89414a136d01fd6a2bb2552eec78e3b86 not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.134112 4759 scope.go:117] "RemoveContainer" containerID="fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.134412 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a\": container with ID starting with fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a not found: ID does not exist" containerID="fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.134465 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a"} err="failed to get container status \"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a\": rpc error: code = NotFound desc = could not find container \"fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a\": container with ID starting with fc8cc1ee3a1455c36ac6242462ba139b30021d128055fdb0e0fd192ad045b59a not found: ID does not exist" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.134486 4759 scope.go:117] "RemoveContainer" containerID="9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489" Nov 25 19:40:21 crc kubenswrapper[4759]: E1125 19:40:21.134735 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489\": container with ID starting with 9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489 not found: ID does not exist" containerID="9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489" Nov 25 19:40:21 crc kubenswrapper[4759]: I1125 19:40:21.134758 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489"} err="failed to get container status \"9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489\": rpc error: code = NotFound desc = could not find container \"9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489\": container with ID starting with 9af1e510b4916d3a86ac00b7be4574ae531f0485a70d7cefc8ac3383fd483489 not found: ID does not exist" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085156 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r42nm"] Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.085856 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085879 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.085900 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085911 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.085923 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085934 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.085952 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085963 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.085983 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.085994 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086009 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086023 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086041 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086052 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086070 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086082 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086096 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086106 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086119 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086129 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086143 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086153 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086167 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086177 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="extract-content" Nov 25 19:40:22 crc kubenswrapper[4759]: E1125 19:40:22.086194 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086204 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="extract-utilities" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086353 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" containerName="marketplace-operator" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086372 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086393 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086407 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.086424 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a685e63a-c301-4410-84f7-6825651ad294" containerName="registry-server" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.087396 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.088621 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r42nm"] Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.089499 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.115557 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eb78bd8-c3b4-4daa-a818-07501af8098f" path="/var/lib/kubelet/pods/3eb78bd8-c3b4-4daa-a818-07501af8098f/volumes" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.116121 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401e8bc0-19b1-440e-8062-298d61c12e37" path="/var/lib/kubelet/pods/401e8bc0-19b1-440e-8062-298d61c12e37/volumes" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.117151 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a685e63a-c301-4410-84f7-6825651ad294" path="/var/lib/kubelet/pods/a685e63a-c301-4410-84f7-6825651ad294/volumes" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.119720 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac427820-58ca-4e4f-b8c6-058084c79617" path="/var/lib/kubelet/pods/ac427820-58ca-4e4f-b8c6-058084c79617/volumes" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.120179 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ded011-6d23-4501-a7ea-f90ed6696cdd" path="/var/lib/kubelet/pods/d0ded011-6d23-4501-a7ea-f90ed6696cdd/volumes" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.181045 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-utilities\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.181187 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-catalog-content\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.181224 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84q9b\" (UniqueName: \"kubernetes.io/projected/bc43e84a-00a8-4187-86ef-232d444a3415-kube-api-access-84q9b\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.224177 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhw6t"] Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.225431 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.227029 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.234143 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhw6t"] Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282343 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-catalog-content\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282383 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-utilities\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282580 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-catalog-content\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282639 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84q9b\" (UniqueName: \"kubernetes.io/projected/bc43e84a-00a8-4187-86ef-232d444a3415-kube-api-access-84q9b\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282704 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr8wv\" (UniqueName: \"kubernetes.io/projected/4570680c-2f2e-44dc-aeec-caf3dd5231ef-kube-api-access-fr8wv\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.282753 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-utilities\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.283135 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-catalog-content\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.284398 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc43e84a-00a8-4187-86ef-232d444a3415-utilities\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.299133 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84q9b\" (UniqueName: \"kubernetes.io/projected/bc43e84a-00a8-4187-86ef-232d444a3415-kube-api-access-84q9b\") pod \"redhat-marketplace-r42nm\" (UID: \"bc43e84a-00a8-4187-86ef-232d444a3415\") " pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.383467 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-catalog-content\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.383511 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-utilities\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.383557 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr8wv\" (UniqueName: \"kubernetes.io/projected/4570680c-2f2e-44dc-aeec-caf3dd5231ef-kube-api-access-fr8wv\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.384148 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-utilities\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.384284 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4570680c-2f2e-44dc-aeec-caf3dd5231ef-catalog-content\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.400066 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr8wv\" (UniqueName: \"kubernetes.io/projected/4570680c-2f2e-44dc-aeec-caf3dd5231ef-kube-api-access-fr8wv\") pod \"redhat-operators-jhw6t\" (UID: \"4570680c-2f2e-44dc-aeec-caf3dd5231ef\") " pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.411832 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.545778 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.584866 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r42nm"] Nov 25 19:40:22 crc kubenswrapper[4759]: W1125 19:40:22.593467 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc43e84a_00a8_4187_86ef_232d444a3415.slice/crio-1ec6b98c59ee88c12aeabf183c5342cc0f3a508dd847149c13816195aa6b9fcc WatchSource:0}: Error finding container 1ec6b98c59ee88c12aeabf183c5342cc0f3a508dd847149c13816195aa6b9fcc: Status 404 returned error can't find the container with id 1ec6b98c59ee88c12aeabf183c5342cc0f3a508dd847149c13816195aa6b9fcc Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.849661 4759 generic.go:334] "Generic (PLEG): container finished" podID="bc43e84a-00a8-4187-86ef-232d444a3415" containerID="3268e5a7e672a6c528be93d91a2401af4a865365b05236edc5523ef7b1e44b8c" exitCode=0 Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.850698 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r42nm" event={"ID":"bc43e84a-00a8-4187-86ef-232d444a3415","Type":"ContainerDied","Data":"3268e5a7e672a6c528be93d91a2401af4a865365b05236edc5523ef7b1e44b8c"} Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.850768 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r42nm" event={"ID":"bc43e84a-00a8-4187-86ef-232d444a3415","Type":"ContainerStarted","Data":"1ec6b98c59ee88c12aeabf183c5342cc0f3a508dd847149c13816195aa6b9fcc"} Nov 25 19:40:22 crc kubenswrapper[4759]: I1125 19:40:22.930742 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhw6t"] Nov 25 19:40:22 crc kubenswrapper[4759]: W1125 19:40:22.938722 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4570680c_2f2e_44dc_aeec_caf3dd5231ef.slice/crio-be4ede9647523312ded5ed811bd0d130e53d9363fe51f55d49a0e40ad3eef93d WatchSource:0}: Error finding container be4ede9647523312ded5ed811bd0d130e53d9363fe51f55d49a0e40ad3eef93d: Status 404 returned error can't find the container with id be4ede9647523312ded5ed811bd0d130e53d9363fe51f55d49a0e40ad3eef93d Nov 25 19:40:23 crc kubenswrapper[4759]: I1125 19:40:23.857170 4759 generic.go:334] "Generic (PLEG): container finished" podID="bc43e84a-00a8-4187-86ef-232d444a3415" containerID="39f618b28c5be8bd5de0acdc99f6b499296a622e568a325eb39d911bb87d524f" exitCode=0 Nov 25 19:40:23 crc kubenswrapper[4759]: I1125 19:40:23.857521 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r42nm" event={"ID":"bc43e84a-00a8-4187-86ef-232d444a3415","Type":"ContainerDied","Data":"39f618b28c5be8bd5de0acdc99f6b499296a622e568a325eb39d911bb87d524f"} Nov 25 19:40:23 crc kubenswrapper[4759]: I1125 19:40:23.860212 4759 generic.go:334] "Generic (PLEG): container finished" podID="4570680c-2f2e-44dc-aeec-caf3dd5231ef" containerID="6f1afed07aead1b5692d7043b32917891ffbce59d4d79cb1613415a832c2f1dc" exitCode=0 Nov 25 19:40:23 crc kubenswrapper[4759]: I1125 19:40:23.860256 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhw6t" event={"ID":"4570680c-2f2e-44dc-aeec-caf3dd5231ef","Type":"ContainerDied","Data":"6f1afed07aead1b5692d7043b32917891ffbce59d4d79cb1613415a832c2f1dc"} Nov 25 19:40:23 crc kubenswrapper[4759]: I1125 19:40:23.860284 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhw6t" event={"ID":"4570680c-2f2e-44dc-aeec-caf3dd5231ef","Type":"ContainerStarted","Data":"be4ede9647523312ded5ed811bd0d130e53d9363fe51f55d49a0e40ad3eef93d"} Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.440908 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ccnmk"] Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.441932 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.444186 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.446772 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ccnmk"] Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.510036 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-catalog-content\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.510130 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8smh4\" (UniqueName: \"kubernetes.io/projected/56bef130-bc8b-49c5-b9e1-c95867515bbc-kube-api-access-8smh4\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.510184 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-utilities\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.611713 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-utilities\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.612463 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-catalog-content\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.612559 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8smh4\" (UniqueName: \"kubernetes.io/projected/56bef130-bc8b-49c5-b9e1-c95867515bbc-kube-api-access-8smh4\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.612651 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-utilities\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.612950 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56bef130-bc8b-49c5-b9e1-c95867515bbc-catalog-content\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.622007 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.623347 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.629535 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.635099 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.635530 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8smh4\" (UniqueName: \"kubernetes.io/projected/56bef130-bc8b-49c5-b9e1-c95867515bbc-kube-api-access-8smh4\") pod \"certified-operators-ccnmk\" (UID: \"56bef130-bc8b-49c5-b9e1-c95867515bbc\") " pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.714314 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.714530 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.714587 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vljtv\" (UniqueName: \"kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.774330 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.816650 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.816745 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vljtv\" (UniqueName: \"kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.816818 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.817765 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.818024 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.847564 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vljtv\" (UniqueName: \"kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv\") pod \"community-operators-5scxm\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.884680 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r42nm" event={"ID":"bc43e84a-00a8-4187-86ef-232d444a3415","Type":"ContainerStarted","Data":"28b99e5f611a289f571ba01e6849e7e51cf7aed682064609672cd828b65a9358"} Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.895360 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.895410 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.895478 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.896131 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.896202 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab" gracePeriod=600 Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.897257 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhw6t" event={"ID":"4570680c-2f2e-44dc-aeec-caf3dd5231ef","Type":"ContainerStarted","Data":"28312b37d9b99f20248b5b03c3014d48893d89d91477f5309ef4c0727ec196df"} Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.911687 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r42nm" podStartSLOduration=1.282916945 podStartE2EDuration="2.911666868s" podCreationTimestamp="2025-11-25 19:40:22 +0000 UTC" firstStartedPulling="2025-11-25 19:40:22.851381129 +0000 UTC m=+413.437328830" lastFinishedPulling="2025-11-25 19:40:24.480131052 +0000 UTC m=+415.066078753" observedRunningTime="2025-11-25 19:40:24.910919707 +0000 UTC m=+415.496867428" watchObservedRunningTime="2025-11-25 19:40:24.911666868 +0000 UTC m=+415.497614569" Nov 25 19:40:24 crc kubenswrapper[4759]: I1125 19:40:24.966816 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.007803 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ccnmk"] Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.158878 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 19:40:25 crc kubenswrapper[4759]: W1125 19:40:25.162564 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a8d174_0d8b_4841_ace2_958422c2d299.slice/crio-fd194fd6ea15f3c70f232ce15673790ce372b8ccd4dd1012672a7429d75bc287 WatchSource:0}: Error finding container fd194fd6ea15f3c70f232ce15673790ce372b8ccd4dd1012672a7429d75bc287: Status 404 returned error can't find the container with id fd194fd6ea15f3c70f232ce15673790ce372b8ccd4dd1012672a7429d75bc287 Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.904036 4759 generic.go:334] "Generic (PLEG): container finished" podID="56bef130-bc8b-49c5-b9e1-c95867515bbc" containerID="7b0504da6113b645c7250f300ef816090bd0105fe2b72b3419a5bd1d132b749d" exitCode=0 Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.904112 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ccnmk" event={"ID":"56bef130-bc8b-49c5-b9e1-c95867515bbc","Type":"ContainerDied","Data":"7b0504da6113b645c7250f300ef816090bd0105fe2b72b3419a5bd1d132b749d"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.904479 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ccnmk" event={"ID":"56bef130-bc8b-49c5-b9e1-c95867515bbc","Type":"ContainerStarted","Data":"825483847dd2afde4bbba6a34b6fcfe03ce63d65d8ce4f805ad47c4617b64da9"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.907372 4759 generic.go:334] "Generic (PLEG): container finished" podID="4570680c-2f2e-44dc-aeec-caf3dd5231ef" containerID="28312b37d9b99f20248b5b03c3014d48893d89d91477f5309ef4c0727ec196df" exitCode=0 Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.907470 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhw6t" event={"ID":"4570680c-2f2e-44dc-aeec-caf3dd5231ef","Type":"ContainerDied","Data":"28312b37d9b99f20248b5b03c3014d48893d89d91477f5309ef4c0727ec196df"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.911639 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab" exitCode=0 Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.911681 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.911720 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.911737 4759 scope.go:117] "RemoveContainer" containerID="41a1fe6ce672763f6975f4e8cf53fba4628b0ea758ee6a65b655f42755526ecf" Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.913102 4759 generic.go:334] "Generic (PLEG): container finished" podID="00a8d174-0d8b-4841-ace2-958422c2d299" containerID="73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f" exitCode=0 Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.914510 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerDied","Data":"73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f"} Nov 25 19:40:25 crc kubenswrapper[4759]: I1125 19:40:25.914539 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerStarted","Data":"fd194fd6ea15f3c70f232ce15673790ce372b8ccd4dd1012672a7429d75bc287"} Nov 25 19:40:26 crc kubenswrapper[4759]: I1125 19:40:26.924682 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ccnmk" event={"ID":"56bef130-bc8b-49c5-b9e1-c95867515bbc","Type":"ContainerStarted","Data":"ace1655f0dfac7f66ea423c1780257aa44975e64feaac25c3f0384fbf14fb704"} Nov 25 19:40:26 crc kubenswrapper[4759]: I1125 19:40:26.929602 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhw6t" event={"ID":"4570680c-2f2e-44dc-aeec-caf3dd5231ef","Type":"ContainerStarted","Data":"0f84e5d2446739b69f170473740bc5fa6caab7107ca90b1944388f5b3259873e"} Nov 25 19:40:26 crc kubenswrapper[4759]: I1125 19:40:26.968978 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhw6t" podStartSLOduration=2.517784092 podStartE2EDuration="4.968951933s" podCreationTimestamp="2025-11-25 19:40:22 +0000 UTC" firstStartedPulling="2025-11-25 19:40:23.861875961 +0000 UTC m=+414.447823662" lastFinishedPulling="2025-11-25 19:40:26.313043802 +0000 UTC m=+416.898991503" observedRunningTime="2025-11-25 19:40:26.96596082 +0000 UTC m=+417.551908521" watchObservedRunningTime="2025-11-25 19:40:26.968951933 +0000 UTC m=+417.554899634" Nov 25 19:40:27 crc kubenswrapper[4759]: I1125 19:40:27.942158 4759 generic.go:334] "Generic (PLEG): container finished" podID="00a8d174-0d8b-4841-ace2-958422c2d299" containerID="2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7" exitCode=0 Nov 25 19:40:27 crc kubenswrapper[4759]: I1125 19:40:27.942234 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerDied","Data":"2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7"} Nov 25 19:40:27 crc kubenswrapper[4759]: I1125 19:40:27.946341 4759 generic.go:334] "Generic (PLEG): container finished" podID="56bef130-bc8b-49c5-b9e1-c95867515bbc" containerID="ace1655f0dfac7f66ea423c1780257aa44975e64feaac25c3f0384fbf14fb704" exitCode=0 Nov 25 19:40:27 crc kubenswrapper[4759]: I1125 19:40:27.947517 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ccnmk" event={"ID":"56bef130-bc8b-49c5-b9e1-c95867515bbc","Type":"ContainerDied","Data":"ace1655f0dfac7f66ea423c1780257aa44975e64feaac25c3f0384fbf14fb704"} Nov 25 19:40:29 crc kubenswrapper[4759]: I1125 19:40:29.957287 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerStarted","Data":"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf"} Nov 25 19:40:29 crc kubenswrapper[4759]: I1125 19:40:29.959335 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ccnmk" event={"ID":"56bef130-bc8b-49c5-b9e1-c95867515bbc","Type":"ContainerStarted","Data":"470dcc157a84c1b701d171e01b3e407ec7f39d97df359fc19225b7655ac5d7f9"} Nov 25 19:40:29 crc kubenswrapper[4759]: I1125 19:40:29.975371 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5scxm" podStartSLOduration=3.557206255 podStartE2EDuration="5.975352805s" podCreationTimestamp="2025-11-25 19:40:24 +0000 UTC" firstStartedPulling="2025-11-25 19:40:25.916134512 +0000 UTC m=+416.502082213" lastFinishedPulling="2025-11-25 19:40:28.334281062 +0000 UTC m=+418.920228763" observedRunningTime="2025-11-25 19:40:29.972695691 +0000 UTC m=+420.558643422" watchObservedRunningTime="2025-11-25 19:40:29.975352805 +0000 UTC m=+420.561300506" Nov 25 19:40:29 crc kubenswrapper[4759]: I1125 19:40:29.993271 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ccnmk" podStartSLOduration=3.098956302 podStartE2EDuration="5.993248079s" podCreationTimestamp="2025-11-25 19:40:24 +0000 UTC" firstStartedPulling="2025-11-25 19:40:25.905843108 +0000 UTC m=+416.491790809" lastFinishedPulling="2025-11-25 19:40:28.800134885 +0000 UTC m=+419.386082586" observedRunningTime="2025-11-25 19:40:29.993050123 +0000 UTC m=+420.578997824" watchObservedRunningTime="2025-11-25 19:40:29.993248079 +0000 UTC m=+420.579195790" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.413313 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.413746 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.454936 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.546808 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.547577 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:32 crc kubenswrapper[4759]: I1125 19:40:32.580410 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:33 crc kubenswrapper[4759]: I1125 19:40:33.013911 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r42nm" Nov 25 19:40:33 crc kubenswrapper[4759]: I1125 19:40:33.030770 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhw6t" Nov 25 19:40:34 crc kubenswrapper[4759]: I1125 19:40:34.775272 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:34 crc kubenswrapper[4759]: I1125 19:40:34.775613 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:34 crc kubenswrapper[4759]: I1125 19:40:34.830205 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:34 crc kubenswrapper[4759]: I1125 19:40:34.967266 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:34 crc kubenswrapper[4759]: I1125 19:40:34.967317 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:35 crc kubenswrapper[4759]: I1125 19:40:35.003591 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:35 crc kubenswrapper[4759]: I1125 19:40:35.023629 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ccnmk" Nov 25 19:40:35 crc kubenswrapper[4759]: I1125 19:40:35.054768 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5scxm" Nov 25 19:40:35 crc kubenswrapper[4759]: I1125 19:40:35.921280 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" podUID="4a6c7e51-7e1e-46d7-9d75-758019749635" containerName="registry" containerID="cri-o://65ac357a459134b8a52666bf82ec6b4aa82597dfeb0d37358f2b78e54919d37c" gracePeriod=30 Nov 25 19:40:37 crc kubenswrapper[4759]: I1125 19:40:37.996562 4759 generic.go:334] "Generic (PLEG): container finished" podID="4a6c7e51-7e1e-46d7-9d75-758019749635" containerID="65ac357a459134b8a52666bf82ec6b4aa82597dfeb0d37358f2b78e54919d37c" exitCode=0 Nov 25 19:40:37 crc kubenswrapper[4759]: I1125 19:40:37.996656 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" event={"ID":"4a6c7e51-7e1e-46d7-9d75-758019749635","Type":"ContainerDied","Data":"65ac357a459134b8a52666bf82ec6b4aa82597dfeb0d37358f2b78e54919d37c"} Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.507626 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608596 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608646 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608678 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608701 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608719 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608761 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbnnd\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608903 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.608954 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls\") pod \"4a6c7e51-7e1e-46d7-9d75-758019749635\" (UID: \"4a6c7e51-7e1e-46d7-9d75-758019749635\") " Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.609611 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.609732 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.620243 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.622205 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.623886 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.637102 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd" (OuterVolumeSpecName: "kube-api-access-qbnnd") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "kube-api-access-qbnnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.644314 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.646528 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4a6c7e51-7e1e-46d7-9d75-758019749635" (UID: "4a6c7e51-7e1e-46d7-9d75-758019749635"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710038 4759 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710076 4759 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a6c7e51-7e1e-46d7-9d75-758019749635-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710087 4759 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a6c7e51-7e1e-46d7-9d75-758019749635-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710095 4759 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a6c7e51-7e1e-46d7-9d75-758019749635-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710103 4759 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710112 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbnnd\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-kube-api-access-qbnnd\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:38 crc kubenswrapper[4759]: I1125 19:40:38.710124 4759 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a6c7e51-7e1e-46d7-9d75-758019749635-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 19:40:39 crc kubenswrapper[4759]: I1125 19:40:39.002414 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" event={"ID":"4a6c7e51-7e1e-46d7-9d75-758019749635","Type":"ContainerDied","Data":"9058f6f9b3fe29d79dc8e21e974de822f495eb50afc689b93add91b455070cc8"} Nov 25 19:40:39 crc kubenswrapper[4759]: I1125 19:40:39.002480 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-khxh2" Nov 25 19:40:39 crc kubenswrapper[4759]: I1125 19:40:39.002498 4759 scope.go:117] "RemoveContainer" containerID="65ac357a459134b8a52666bf82ec6b4aa82597dfeb0d37358f2b78e54919d37c" Nov 25 19:40:39 crc kubenswrapper[4759]: I1125 19:40:39.025691 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:40:39 crc kubenswrapper[4759]: I1125 19:40:39.028996 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-khxh2"] Nov 25 19:40:40 crc kubenswrapper[4759]: I1125 19:40:40.122661 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a6c7e51-7e1e-46d7-9d75-758019749635" path="/var/lib/kubelet/pods/4a6c7e51-7e1e-46d7-9d75-758019749635/volumes" Nov 25 19:42:54 crc kubenswrapper[4759]: I1125 19:42:54.895776 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:42:54 crc kubenswrapper[4759]: I1125 19:42:54.897619 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:43:24 crc kubenswrapper[4759]: I1125 19:43:24.895371 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:43:24 crc kubenswrapper[4759]: I1125 19:43:24.896022 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:43:54 crc kubenswrapper[4759]: I1125 19:43:54.895134 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:43:54 crc kubenswrapper[4759]: I1125 19:43:54.895712 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:43:54 crc kubenswrapper[4759]: I1125 19:43:54.895772 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:43:54 crc kubenswrapper[4759]: I1125 19:43:54.896373 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:43:54 crc kubenswrapper[4759]: I1125 19:43:54.896474 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b" gracePeriod=600 Nov 25 19:43:55 crc kubenswrapper[4759]: I1125 19:43:55.100154 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b" exitCode=0 Nov 25 19:43:55 crc kubenswrapper[4759]: I1125 19:43:55.100200 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b"} Nov 25 19:43:55 crc kubenswrapper[4759]: I1125 19:43:55.100294 4759 scope.go:117] "RemoveContainer" containerID="41bb8a18b4ad55a5c36bc99417b9ee3006209e50ea54a05deec69e6ac3b611ab" Nov 25 19:43:56 crc kubenswrapper[4759]: I1125 19:43:56.111955 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18"} Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.180405 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8"] Nov 25 19:45:00 crc kubenswrapper[4759]: E1125 19:45:00.182121 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6c7e51-7e1e-46d7-9d75-758019749635" containerName="registry" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.182141 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6c7e51-7e1e-46d7-9d75-758019749635" containerName="registry" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.182265 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a6c7e51-7e1e-46d7-9d75-758019749635" containerName="registry" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.182742 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.187085 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8"] Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.187907 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.188136 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.364480 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.364869 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.364918 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxd68\" (UniqueName: \"kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.466181 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.466252 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.466287 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxd68\" (UniqueName: \"kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.468206 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.475023 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.482633 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxd68\" (UniqueName: \"kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68\") pod \"collect-profiles-29401665-fgrr8\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.534280 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:00 crc kubenswrapper[4759]: I1125 19:45:00.941373 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8"] Nov 25 19:45:01 crc kubenswrapper[4759]: I1125 19:45:01.494300 4759 generic.go:334] "Generic (PLEG): container finished" podID="5834c2b9-f48f-438b-bdfc-01913c6d14d2" containerID="8de671a7b1a44e35f7a9cf2a5248d3010ea6b12a055e54ab0b21f213c3f8a5d0" exitCode=0 Nov 25 19:45:01 crc kubenswrapper[4759]: I1125 19:45:01.494354 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" event={"ID":"5834c2b9-f48f-438b-bdfc-01913c6d14d2","Type":"ContainerDied","Data":"8de671a7b1a44e35f7a9cf2a5248d3010ea6b12a055e54ab0b21f213c3f8a5d0"} Nov 25 19:45:01 crc kubenswrapper[4759]: I1125 19:45:01.494763 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" event={"ID":"5834c2b9-f48f-438b-bdfc-01913c6d14d2","Type":"ContainerStarted","Data":"78743d28c805063724c0d8f928b2be36d987853923ae8cbc4ee6c7a53585fec1"} Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.697128 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.896100 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxd68\" (UniqueName: \"kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68\") pod \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.896167 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume\") pod \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.896204 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume\") pod \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\" (UID: \"5834c2b9-f48f-438b-bdfc-01913c6d14d2\") " Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.896761 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "5834c2b9-f48f-438b-bdfc-01913c6d14d2" (UID: "5834c2b9-f48f-438b-bdfc-01913c6d14d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.901550 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68" (OuterVolumeSpecName: "kube-api-access-qxd68") pod "5834c2b9-f48f-438b-bdfc-01913c6d14d2" (UID: "5834c2b9-f48f-438b-bdfc-01913c6d14d2"). InnerVolumeSpecName "kube-api-access-qxd68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.901864 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5834c2b9-f48f-438b-bdfc-01913c6d14d2" (UID: "5834c2b9-f48f-438b-bdfc-01913c6d14d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.997645 4759 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5834c2b9-f48f-438b-bdfc-01913c6d14d2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.997677 4759 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5834c2b9-f48f-438b-bdfc-01913c6d14d2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:02 crc kubenswrapper[4759]: I1125 19:45:02.997687 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxd68\" (UniqueName: \"kubernetes.io/projected/5834c2b9-f48f-438b-bdfc-01913c6d14d2-kube-api-access-qxd68\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:03 crc kubenswrapper[4759]: I1125 19:45:03.511268 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" event={"ID":"5834c2b9-f48f-438b-bdfc-01913c6d14d2","Type":"ContainerDied","Data":"78743d28c805063724c0d8f928b2be36d987853923ae8cbc4ee6c7a53585fec1"} Nov 25 19:45:03 crc kubenswrapper[4759]: I1125 19:45:03.511336 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78743d28c805063724c0d8f928b2be36d987853923ae8cbc4ee6c7a53585fec1" Nov 25 19:45:03 crc kubenswrapper[4759]: I1125 19:45:03.511514 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401665-fgrr8" Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.667046 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-99nmn"] Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668514 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-controller" containerID="cri-o://788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668610 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="nbdb" containerID="cri-o://8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668725 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="northd" containerID="cri-o://f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668813 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-node" containerID="cri-o://24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668864 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668789 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="sbdb" containerID="cri-o://fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.668893 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-acl-logging" containerID="cri-o://0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.702437 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" containerID="cri-o://4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" gracePeriod=30 Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.973421 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/4.log" Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.974241 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/3.log" Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.975660 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovn-acl-logging/0.log" Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.976006 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovn-controller/0.log" Nov 25 19:45:32 crc kubenswrapper[4759]: I1125 19:45:32.976355 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017533 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rhmnt"] Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017769 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kubecfg-setup" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017780 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kubecfg-setup" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017790 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="nbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017796 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="nbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017803 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017809 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017816 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017821 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017829 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017834 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017841 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-acl-logging" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017847 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-acl-logging" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017854 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-node" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017859 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-node" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017870 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="northd" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017875 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="northd" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017881 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017887 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017894 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5834c2b9-f48f-438b-bdfc-01913c6d14d2" containerName="collect-profiles" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017901 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="5834c2b9-f48f-438b-bdfc-01913c6d14d2" containerName="collect-profiles" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017908 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017914 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017921 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017927 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.017939 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="sbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.017944 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="sbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018020 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018028 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-acl-logging" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018035 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovn-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018043 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018053 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018059 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="northd" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018071 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="5834c2b9-f48f-438b-bdfc-01913c6d14d2" containerName="collect-profiles" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018077 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018084 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="sbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018091 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018097 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="kube-rbac-proxy-node" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018107 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="nbdb" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.018185 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018192 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.018285 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerName="ovnkube-controller" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.019910 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169282 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169332 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169348 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169370 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169387 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169405 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169428 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lr5j\" (UniqueName: \"kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169494 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169528 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169576 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169600 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169623 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169363 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169657 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169685 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169399 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169727 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169419 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169462 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169626 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket" (OuterVolumeSpecName: "log-socket") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169675 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169680 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169764 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169702 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log" (OuterVolumeSpecName: "node-log") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169704 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169697 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169709 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169721 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash" (OuterVolumeSpecName: "host-slash") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169725 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169863 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169888 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169918 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169950 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides\") pod \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\" (UID: \"d8b6037c-81e1-4fe8-9216-5b1863a029f1\") " Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.169952 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170359 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-ovn\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170422 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170438 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-systemd-units\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170489 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170513 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170545 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-node-log\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170565 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-bin\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170588 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170616 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-var-lib-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170649 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-etc-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170671 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovn-node-metrics-cert\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170732 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-netd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170772 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-systemd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170787 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170792 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-config\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170863 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-kubelet\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170898 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-slash\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170916 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-script-lib\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170938 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr5qm\" (UniqueName: \"kubernetes.io/projected/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-kube-api-access-mr5qm\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170958 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-log-socket\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170973 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-netns\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.170987 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-env-overrides\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171010 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171078 4759 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171092 4759 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171104 4759 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171114 4759 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171125 4759 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8b6037c-81e1-4fe8-9216-5b1863a029f1-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171136 4759 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171147 4759 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171158 4759 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171168 4759 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171178 4759 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171189 4759 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171199 4759 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171209 4759 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171218 4759 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171227 4759 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171236 4759 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.171246 4759 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.175458 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j" (OuterVolumeSpecName: "kube-api-access-8lr5j") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "kube-api-access-8lr5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.175695 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.183456 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d8b6037c-81e1-4fe8-9216-5b1863a029f1" (UID: "d8b6037c-81e1-4fe8-9216-5b1863a029f1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272515 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272596 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-node-log\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272617 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272638 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-bin\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272691 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272703 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-bin\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272736 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-var-lib-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272692 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-node-log\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272778 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-etc-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272791 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272836 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-etc-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272803 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovn-node-metrics-cert\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272804 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-var-lib-openvswitch\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272913 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-netd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272949 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-cni-netd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.272978 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-systemd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273007 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-config\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273040 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-kubelet\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273041 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-systemd\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273060 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-slash\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273080 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-script-lib\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273103 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr5qm\" (UniqueName: \"kubernetes.io/projected/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-kube-api-access-mr5qm\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273107 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-kubelet\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273126 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-slash\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273153 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-log-socket\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273176 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-netns\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273193 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-env-overrides\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273213 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273273 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-run-netns\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273280 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-ovn\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273311 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-log-socket\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273343 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273358 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-systemd-units\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273381 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-run-ovn\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273472 4759 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8b6037c-81e1-4fe8-9216-5b1863a029f1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273491 4759 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8b6037c-81e1-4fe8-9216-5b1863a029f1-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273531 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-systemd-units\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273580 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lr5j\" (UniqueName: \"kubernetes.io/projected/d8b6037c-81e1-4fe8-9216-5b1863a029f1-kube-api-access-8lr5j\") on node \"crc\" DevicePath \"\"" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.273934 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-config\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.274536 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-env-overrides\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.274579 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovnkube-script-lib\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.276396 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-ovn-node-metrics-cert\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.290142 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr5qm\" (UniqueName: \"kubernetes.io/projected/3e1aae3d-97d7-43cf-8031-4d7ab63bca86-kube-api-access-mr5qm\") pod \"ovnkube-node-rhmnt\" (UID: \"3e1aae3d-97d7-43cf-8031-4d7ab63bca86\") " pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.340382 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.696922 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/4.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.697581 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovnkube-controller/3.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.700098 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovn-acl-logging/0.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.700931 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-99nmn_d8b6037c-81e1-4fe8-9216-5b1863a029f1/ovn-controller/0.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701318 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" exitCode=2 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701351 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701390 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701400 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701409 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701417 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701412 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701425 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" exitCode=143 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701479 4759 generic.go:334] "Generic (PLEG): container finished" podID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" exitCode=143 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701427 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701482 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701553 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701571 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701588 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701603 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701615 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701626 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701632 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701637 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701643 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701649 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701654 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701659 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701664 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701672 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701679 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701686 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701692 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701696 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701701 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701706 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701711 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701716 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701721 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701726 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701733 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701739 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701746 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701751 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701756 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701762 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701766 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701497 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701771 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701866 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701878 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701886 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701898 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-99nmn" event={"ID":"d8b6037c-81e1-4fe8-9216-5b1863a029f1","Type":"ContainerDied","Data":"cd42469ea6973db6d014ce9669d6690b8067361de911ac8e65bdab3b5cd8fe84"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701915 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701924 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701931 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701939 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701945 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701951 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701958 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701964 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701972 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.701978 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.704118 4759 generic.go:334] "Generic (PLEG): container finished" podID="3e1aae3d-97d7-43cf-8031-4d7ab63bca86" containerID="b4b2f898e9d0bae1f8dadccb5a423bdd63a2fbedc03557c90e0bf6fe5e91a72f" exitCode=0 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.704134 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerDied","Data":"b4b2f898e9d0bae1f8dadccb5a423bdd63a2fbedc03557c90e0bf6fe5e91a72f"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.704158 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"4224211c677042b25782304a8a6ad0c8901147561451b2a1db9a345fd44de6ac"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.707470 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/2.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.707893 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/1.log" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.707934 4759 generic.go:334] "Generic (PLEG): container finished" podID="67e7b861-9f72-41ba-ab98-35e0a2bdbe39" containerID="f203afb0342346ff4202e818d51b048ac2643cdbdd1aeb099a142dc45aab802c" exitCode=2 Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.707960 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerDied","Data":"f203afb0342346ff4202e818d51b048ac2643cdbdd1aeb099a142dc45aab802c"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.707982 4759 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353"} Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.708333 4759 scope.go:117] "RemoveContainer" containerID="f203afb0342346ff4202e818d51b048ac2643cdbdd1aeb099a142dc45aab802c" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.708539 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vz6pj_openshift-multus(67e7b861-9f72-41ba-ab98-35e0a2bdbe39)\"" pod="openshift-multus/multus-vz6pj" podUID="67e7b861-9f72-41ba-ab98-35e0a2bdbe39" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.722047 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.750862 4759 scope.go:117] "RemoveContainer" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.774359 4759 scope.go:117] "RemoveContainer" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.776475 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-99nmn"] Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.787539 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-99nmn"] Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.832666 4759 scope.go:117] "RemoveContainer" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.847980 4759 scope.go:117] "RemoveContainer" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.864981 4759 scope.go:117] "RemoveContainer" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.885052 4759 scope.go:117] "RemoveContainer" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.908537 4759 scope.go:117] "RemoveContainer" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.937601 4759 scope.go:117] "RemoveContainer" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.956607 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.957132 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.957173 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} err="failed to get container status \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.957195 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.957713 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": container with ID starting with c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59 not found: ID does not exist" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.957748 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} err="failed to get container status \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": rpc error: code = NotFound desc = could not find container \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": container with ID starting with c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.957765 4759 scope.go:117] "RemoveContainer" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.958567 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": container with ID starting with fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640 not found: ID does not exist" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.958593 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} err="failed to get container status \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": rpc error: code = NotFound desc = could not find container \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": container with ID starting with fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.958607 4759 scope.go:117] "RemoveContainer" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.959066 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": container with ID starting with 8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7 not found: ID does not exist" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959089 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} err="failed to get container status \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": rpc error: code = NotFound desc = could not find container \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": container with ID starting with 8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959106 4759 scope.go:117] "RemoveContainer" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.959512 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": container with ID starting with f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc not found: ID does not exist" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959533 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} err="failed to get container status \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": rpc error: code = NotFound desc = could not find container \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": container with ID starting with f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959561 4759 scope.go:117] "RemoveContainer" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.959893 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": container with ID starting with 017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a not found: ID does not exist" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959935 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} err="failed to get container status \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": rpc error: code = NotFound desc = could not find container \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": container with ID starting with 017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.959966 4759 scope.go:117] "RemoveContainer" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.960389 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": container with ID starting with 24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531 not found: ID does not exist" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.960418 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} err="failed to get container status \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": rpc error: code = NotFound desc = could not find container \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": container with ID starting with 24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.960437 4759 scope.go:117] "RemoveContainer" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.960848 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": container with ID starting with 0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f not found: ID does not exist" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.960877 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} err="failed to get container status \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": rpc error: code = NotFound desc = could not find container \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": container with ID starting with 0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.960893 4759 scope.go:117] "RemoveContainer" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.961262 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": container with ID starting with 788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f not found: ID does not exist" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.961366 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} err="failed to get container status \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": rpc error: code = NotFound desc = could not find container \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": container with ID starting with 788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.961401 4759 scope.go:117] "RemoveContainer" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: E1125 19:45:33.961762 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": container with ID starting with b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526 not found: ID does not exist" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.961787 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} err="failed to get container status \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": rpc error: code = NotFound desc = could not find container \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": container with ID starting with b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.961800 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962077 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} err="failed to get container status \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962095 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962368 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} err="failed to get container status \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": rpc error: code = NotFound desc = could not find container \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": container with ID starting with c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962389 4759 scope.go:117] "RemoveContainer" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962669 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} err="failed to get container status \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": rpc error: code = NotFound desc = could not find container \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": container with ID starting with fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962696 4759 scope.go:117] "RemoveContainer" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962918 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} err="failed to get container status \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": rpc error: code = NotFound desc = could not find container \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": container with ID starting with 8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.962942 4759 scope.go:117] "RemoveContainer" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.963132 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} err="failed to get container status \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": rpc error: code = NotFound desc = could not find container \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": container with ID starting with f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.963152 4759 scope.go:117] "RemoveContainer" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.963363 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} err="failed to get container status \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": rpc error: code = NotFound desc = could not find container \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": container with ID starting with 017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.963388 4759 scope.go:117] "RemoveContainer" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.964042 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} err="failed to get container status \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": rpc error: code = NotFound desc = could not find container \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": container with ID starting with 24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.964065 4759 scope.go:117] "RemoveContainer" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.964277 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} err="failed to get container status \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": rpc error: code = NotFound desc = could not find container \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": container with ID starting with 0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.964298 4759 scope.go:117] "RemoveContainer" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965051 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} err="failed to get container status \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": rpc error: code = NotFound desc = could not find container \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": container with ID starting with 788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965074 4759 scope.go:117] "RemoveContainer" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965333 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} err="failed to get container status \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": rpc error: code = NotFound desc = could not find container \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": container with ID starting with b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965355 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965613 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} err="failed to get container status \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.965634 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966001 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} err="failed to get container status \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": rpc error: code = NotFound desc = could not find container \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": container with ID starting with c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966023 4759 scope.go:117] "RemoveContainer" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966284 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} err="failed to get container status \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": rpc error: code = NotFound desc = could not find container \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": container with ID starting with fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966307 4759 scope.go:117] "RemoveContainer" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966567 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} err="failed to get container status \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": rpc error: code = NotFound desc = could not find container \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": container with ID starting with 8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966587 4759 scope.go:117] "RemoveContainer" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966935 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} err="failed to get container status \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": rpc error: code = NotFound desc = could not find container \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": container with ID starting with f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.966955 4759 scope.go:117] "RemoveContainer" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.967234 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} err="failed to get container status \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": rpc error: code = NotFound desc = could not find container \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": container with ID starting with 017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.967252 4759 scope.go:117] "RemoveContainer" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.967464 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} err="failed to get container status \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": rpc error: code = NotFound desc = could not find container \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": container with ID starting with 24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.967486 4759 scope.go:117] "RemoveContainer" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.967958 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} err="failed to get container status \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": rpc error: code = NotFound desc = could not find container \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": container with ID starting with 0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.968079 4759 scope.go:117] "RemoveContainer" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.968645 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} err="failed to get container status \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": rpc error: code = NotFound desc = could not find container \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": container with ID starting with 788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.968667 4759 scope.go:117] "RemoveContainer" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969012 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} err="failed to get container status \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": rpc error: code = NotFound desc = could not find container \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": container with ID starting with b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969032 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969482 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} err="failed to get container status \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969502 4759 scope.go:117] "RemoveContainer" containerID="c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969917 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59"} err="failed to get container status \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": rpc error: code = NotFound desc = could not find container \"c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59\": container with ID starting with c3a9ae3dc3049632e282c3a9e194affbf83e6881dcf5c2149704b94fde177f59 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.969933 4759 scope.go:117] "RemoveContainer" containerID="fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.970283 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640"} err="failed to get container status \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": rpc error: code = NotFound desc = could not find container \"fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640\": container with ID starting with fa350ffbe2d5b00a432847517494b3e2a59e868059db5b37a251191b1aba0640 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.970347 4759 scope.go:117] "RemoveContainer" containerID="8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.970712 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7"} err="failed to get container status \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": rpc error: code = NotFound desc = could not find container \"8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7\": container with ID starting with 8a5608f72f37cd79ea43c381bf1b878c1fcb05a78ccc36c0cd4a7f428e7036a7 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.970739 4759 scope.go:117] "RemoveContainer" containerID="f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.971126 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc"} err="failed to get container status \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": rpc error: code = NotFound desc = could not find container \"f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc\": container with ID starting with f60b56d31eff0fb68a1196e8cba5d8abc2842b9b03cc1301471e566101362fbc not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.971153 4759 scope.go:117] "RemoveContainer" containerID="017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.971464 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a"} err="failed to get container status \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": rpc error: code = NotFound desc = could not find container \"017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a\": container with ID starting with 017a5343b3be2f2478de9bcb6b99fe08a750c73e489e52ffb6bc8028b4e8781a not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.971485 4759 scope.go:117] "RemoveContainer" containerID="24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.972427 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531"} err="failed to get container status \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": rpc error: code = NotFound desc = could not find container \"24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531\": container with ID starting with 24e8d0f37c94567231a2b3f37095c27c6a543dd961b168e4a34eab151ce9f531 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.972466 4759 scope.go:117] "RemoveContainer" containerID="0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.972790 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f"} err="failed to get container status \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": rpc error: code = NotFound desc = could not find container \"0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f\": container with ID starting with 0bffe1e4470434cc0951b279dd00dcbc84c5599a1375e0cadb4bc45aed24f28f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.972811 4759 scope.go:117] "RemoveContainer" containerID="788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.973053 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f"} err="failed to get container status \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": rpc error: code = NotFound desc = could not find container \"788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f\": container with ID starting with 788b7c8f97fe298f5e8c8d1df4efd79c8bfffcf7800a2e776cf81e964e7f812f not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.973074 4759 scope.go:117] "RemoveContainer" containerID="b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.973423 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526"} err="failed to get container status \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": rpc error: code = NotFound desc = could not find container \"b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526\": container with ID starting with b63a9be2e34c415c50c8f662be78f01828bd8fe6bc227f16bab9ae020a117526 not found: ID does not exist" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.973524 4759 scope.go:117] "RemoveContainer" containerID="4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b" Nov 25 19:45:33 crc kubenswrapper[4759]: I1125 19:45:33.973846 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b"} err="failed to get container status \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": rpc error: code = NotFound desc = could not find container \"4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b\": container with ID starting with 4c37eaef7e3ebc8e32e7f85ed3e972dbcc333557becef400bee4e97f489dc53b not found: ID does not exist" Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.113590 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8b6037c-81e1-4fe8-9216-5b1863a029f1" path="/var/lib/kubelet/pods/d8b6037c-81e1-4fe8-9216-5b1863a029f1/volumes" Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.720969 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"601e4b9d52f03a945a65d52876f33a8f5c54b6971eb01aaef728b64d1511b908"} Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.721555 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"0c7245b57e4ad9d20ddb037b8603bf2ed84c8ab306132a9f8fabd115ee542550"} Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.721571 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"97e7c29ecdd4aff41cbfc47d355086e8a34c0a83edfd259079c70d3082a83582"} Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.721605 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"ac1eaaab38787215dc08adbf55f644b176789f2493a3b5d4488615957e036270"} Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.721622 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"a738988fd7a67c419637a024558a22692d2040d83e35ed0c1fa084f178b3c5cf"} Nov 25 19:45:34 crc kubenswrapper[4759]: I1125 19:45:34.721634 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"4fb493f2f502415d7db97d1c7547d1c88bcb27e758e4aa9c3d6e8542352678ce"} Nov 25 19:45:37 crc kubenswrapper[4759]: I1125 19:45:37.747824 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"a2292806a066302836e50b681bc427e3d231ef09ecdd398fb71b1629cbb2395c"} Nov 25 19:45:39 crc kubenswrapper[4759]: I1125 19:45:39.771584 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" event={"ID":"3e1aae3d-97d7-43cf-8031-4d7ab63bca86","Type":"ContainerStarted","Data":"00b6028838933861cd18ee8894524d3ebeaf0607c27f2eda935479bbbc0f6717"} Nov 25 19:45:39 crc kubenswrapper[4759]: I1125 19:45:39.800342 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" podStartSLOduration=6.800323156 podStartE2EDuration="6.800323156s" podCreationTimestamp="2025-11-25 19:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:45:39.798932377 +0000 UTC m=+730.384880098" watchObservedRunningTime="2025-11-25 19:45:39.800323156 +0000 UTC m=+730.386270857" Nov 25 19:45:40 crc kubenswrapper[4759]: I1125 19:45:40.776146 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:40 crc kubenswrapper[4759]: I1125 19:45:40.776552 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:40 crc kubenswrapper[4759]: I1125 19:45:40.776573 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:40 crc kubenswrapper[4759]: I1125 19:45:40.811898 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:40 crc kubenswrapper[4759]: I1125 19:45:40.817231 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:45:49 crc kubenswrapper[4759]: I1125 19:45:49.107405 4759 scope.go:117] "RemoveContainer" containerID="f203afb0342346ff4202e818d51b048ac2643cdbdd1aeb099a142dc45aab802c" Nov 25 19:45:49 crc kubenswrapper[4759]: I1125 19:45:49.824675 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/2.log" Nov 25 19:45:49 crc kubenswrapper[4759]: I1125 19:45:49.825575 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/1.log" Nov 25 19:45:49 crc kubenswrapper[4759]: I1125 19:45:49.825612 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vz6pj" event={"ID":"67e7b861-9f72-41ba-ab98-35e0a2bdbe39","Type":"ContainerStarted","Data":"a388d38b7bc2442730d50d89ef68aab322715288127b3c6a68327425ce1e9011"} Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.316501 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq"] Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.318033 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.320743 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.330547 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq"] Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.393550 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.393621 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm7dp\" (UniqueName: \"kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.393714 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.494921 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.494982 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.495026 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm7dp\" (UniqueName: \"kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.495567 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.495571 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.512527 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm7dp\" (UniqueName: \"kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.646115 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.833068 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq"] Nov 25 19:45:57 crc kubenswrapper[4759]: W1125 19:45:57.838620 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4147e62_c53d_46d2_89f2_bc1ccb3859e7.slice/crio-059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f WatchSource:0}: Error finding container 059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f: Status 404 returned error can't find the container with id 059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f Nov 25 19:45:57 crc kubenswrapper[4759]: I1125 19:45:57.866028 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" event={"ID":"a4147e62-c53d-46d2-89f2-bc1ccb3859e7","Type":"ContainerStarted","Data":"059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f"} Nov 25 19:45:58 crc kubenswrapper[4759]: I1125 19:45:58.877991 4759 generic.go:334] "Generic (PLEG): container finished" podID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerID="bc46c1e88772d68391cae17dfb77b9a89c8fc7d2a4b5f9d47ad9a31939779a7c" exitCode=0 Nov 25 19:45:58 crc kubenswrapper[4759]: I1125 19:45:58.878507 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" event={"ID":"a4147e62-c53d-46d2-89f2-bc1ccb3859e7","Type":"ContainerDied","Data":"bc46c1e88772d68391cae17dfb77b9a89c8fc7d2a4b5f9d47ad9a31939779a7c"} Nov 25 19:45:58 crc kubenswrapper[4759]: I1125 19:45:58.883851 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 19:46:00 crc kubenswrapper[4759]: I1125 19:46:00.891623 4759 generic.go:334] "Generic (PLEG): container finished" podID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerID="e12eae45d7877eb87b27c9d7987ba91cf1137e4759e7b8df8e0537711ad6cee5" exitCode=0 Nov 25 19:46:00 crc kubenswrapper[4759]: I1125 19:46:00.891742 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" event={"ID":"a4147e62-c53d-46d2-89f2-bc1ccb3859e7","Type":"ContainerDied","Data":"e12eae45d7877eb87b27c9d7987ba91cf1137e4759e7b8df8e0537711ad6cee5"} Nov 25 19:46:01 crc kubenswrapper[4759]: I1125 19:46:01.900589 4759 generic.go:334] "Generic (PLEG): container finished" podID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerID="900aafb3c914aec1f45732992aae762e6db6107d3df127bdeb0a3ef5af243992" exitCode=0 Nov 25 19:46:01 crc kubenswrapper[4759]: I1125 19:46:01.900642 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" event={"ID":"a4147e62-c53d-46d2-89f2-bc1ccb3859e7","Type":"ContainerDied","Data":"900aafb3c914aec1f45732992aae762e6db6107d3df127bdeb0a3ef5af243992"} Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.135022 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.266920 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util\") pod \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.266958 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle\") pod \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.267037 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm7dp\" (UniqueName: \"kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp\") pod \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\" (UID: \"a4147e62-c53d-46d2-89f2-bc1ccb3859e7\") " Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.268151 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle" (OuterVolumeSpecName: "bundle") pod "a4147e62-c53d-46d2-89f2-bc1ccb3859e7" (UID: "a4147e62-c53d-46d2-89f2-bc1ccb3859e7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.272551 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp" (OuterVolumeSpecName: "kube-api-access-pm7dp") pod "a4147e62-c53d-46d2-89f2-bc1ccb3859e7" (UID: "a4147e62-c53d-46d2-89f2-bc1ccb3859e7"). InnerVolumeSpecName "kube-api-access-pm7dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.280967 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util" (OuterVolumeSpecName: "util") pod "a4147e62-c53d-46d2-89f2-bc1ccb3859e7" (UID: "a4147e62-c53d-46d2-89f2-bc1ccb3859e7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.359659 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rhmnt" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.368120 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.368278 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.368296 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm7dp\" (UniqueName: \"kubernetes.io/projected/a4147e62-c53d-46d2-89f2-bc1ccb3859e7-kube-api-access-pm7dp\") on node \"crc\" DevicePath \"\"" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.913273 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" event={"ID":"a4147e62-c53d-46d2-89f2-bc1ccb3859e7","Type":"ContainerDied","Data":"059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f"} Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.913636 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="059b00f683a0e1d233610038c06ccbc07326cba67fcfcb661f3e60574a9b062f" Nov 25 19:46:03 crc kubenswrapper[4759]: I1125 19:46:03.913308 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.831396 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-747d477645-tgk7n"] Nov 25 19:46:12 crc kubenswrapper[4759]: E1125 19:46:12.832117 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="pull" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.832130 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="pull" Nov 25 19:46:12 crc kubenswrapper[4759]: E1125 19:46:12.832142 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="extract" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.832148 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="extract" Nov 25 19:46:12 crc kubenswrapper[4759]: E1125 19:46:12.832160 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="util" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.832166 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="util" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.832246 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4147e62-c53d-46d2-89f2-bc1ccb3859e7" containerName="extract" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.832608 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.834731 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.835539 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.835828 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.835857 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sm2jq" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.847057 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.848658 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-747d477645-tgk7n"] Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.981912 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pzs2\" (UniqueName: \"kubernetes.io/projected/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-kube-api-access-8pzs2\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.981977 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-apiservice-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:12 crc kubenswrapper[4759]: I1125 19:46:12.982024 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-webhook-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.082696 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-apiservice-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.082763 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-webhook-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.082821 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pzs2\" (UniqueName: \"kubernetes.io/projected/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-kube-api-access-8pzs2\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.088694 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-apiservice-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.091549 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-webhook-cert\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.097877 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pzs2\" (UniqueName: \"kubernetes.io/projected/29fe3fc7-edb3-4ffd-b213-b53a2a055a1b-kube-api-access-8pzs2\") pod \"metallb-operator-controller-manager-747d477645-tgk7n\" (UID: \"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b\") " pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.150388 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.274503 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n"] Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.275187 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.279642 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-49kmt" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.279808 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.279864 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.287156 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-webhook-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.287251 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-apiservice-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.287285 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgbl\" (UniqueName: \"kubernetes.io/projected/94d17030-8068-486d-906f-558adb93cdfb-kube-api-access-qjgbl\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.288474 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n"] Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.387978 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-webhook-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.388040 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-apiservice-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.388142 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgbl\" (UniqueName: \"kubernetes.io/projected/94d17030-8068-486d-906f-558adb93cdfb-kube-api-access-qjgbl\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.397604 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-apiservice-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.419766 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d17030-8068-486d-906f-558adb93cdfb-webhook-cert\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.426108 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-747d477645-tgk7n"] Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.434318 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgbl\" (UniqueName: \"kubernetes.io/projected/94d17030-8068-486d-906f-558adb93cdfb-kube-api-access-qjgbl\") pod \"metallb-operator-webhook-server-84b47d7745-jnp5n\" (UID: \"94d17030-8068-486d-906f-558adb93cdfb\") " pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.474221 4759 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.645031 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.857288 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n"] Nov 25 19:46:13 crc kubenswrapper[4759]: W1125 19:46:13.870989 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94d17030_8068_486d_906f_558adb93cdfb.slice/crio-6849d41d0d098b14ebb4d80c8beebb57400fe344ebd84f35258505da859259e8 WatchSource:0}: Error finding container 6849d41d0d098b14ebb4d80c8beebb57400fe344ebd84f35258505da859259e8: Status 404 returned error can't find the container with id 6849d41d0d098b14ebb4d80c8beebb57400fe344ebd84f35258505da859259e8 Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.956642 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" event={"ID":"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b","Type":"ContainerStarted","Data":"257bff0a8bf62e383fb2204e18a4725f08bd3f3086fe90391b772c50140b1cc5"} Nov 25 19:46:13 crc kubenswrapper[4759]: I1125 19:46:13.957553 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" event={"ID":"94d17030-8068-486d-906f-558adb93cdfb","Type":"ContainerStarted","Data":"6849d41d0d098b14ebb4d80c8beebb57400fe344ebd84f35258505da859259e8"} Nov 25 19:46:18 crc kubenswrapper[4759]: I1125 19:46:18.993327 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" event={"ID":"29fe3fc7-edb3-4ffd-b213-b53a2a055a1b","Type":"ContainerStarted","Data":"a2e1cdfbfbcd21d912fdfcaffa434947d47f308b7bd8def7449880ab6045eb9f"} Nov 25 19:46:18 crc kubenswrapper[4759]: I1125 19:46:18.994172 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:18 crc kubenswrapper[4759]: I1125 19:46:18.995867 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" event={"ID":"94d17030-8068-486d-906f-558adb93cdfb","Type":"ContainerStarted","Data":"328fb651c7c5ae5617873ae70c056fae217e13ac5f9963deab14e0e65bd36ef7"} Nov 25 19:46:18 crc kubenswrapper[4759]: I1125 19:46:18.996005 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:19 crc kubenswrapper[4759]: I1125 19:46:19.026435 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" podStartSLOduration=2.3839121309999998 podStartE2EDuration="7.026411282s" podCreationTimestamp="2025-11-25 19:46:12 +0000 UTC" firstStartedPulling="2025-11-25 19:46:13.457698216 +0000 UTC m=+764.043645917" lastFinishedPulling="2025-11-25 19:46:18.100197367 +0000 UTC m=+768.686145068" observedRunningTime="2025-11-25 19:46:19.023570632 +0000 UTC m=+769.609518373" watchObservedRunningTime="2025-11-25 19:46:19.026411282 +0000 UTC m=+769.612359003" Nov 25 19:46:19 crc kubenswrapper[4759]: I1125 19:46:19.047308 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" podStartSLOduration=1.7917592500000001 podStartE2EDuration="6.047288008s" podCreationTimestamp="2025-11-25 19:46:13 +0000 UTC" firstStartedPulling="2025-11-25 19:46:13.874838707 +0000 UTC m=+764.460786408" lastFinishedPulling="2025-11-25 19:46:18.130367465 +0000 UTC m=+768.716315166" observedRunningTime="2025-11-25 19:46:19.046467005 +0000 UTC m=+769.632414746" watchObservedRunningTime="2025-11-25 19:46:19.047288008 +0000 UTC m=+769.633235709" Nov 25 19:46:24 crc kubenswrapper[4759]: I1125 19:46:24.895995 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:46:24 crc kubenswrapper[4759]: I1125 19:46:24.896377 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:46:30 crc kubenswrapper[4759]: I1125 19:46:30.375975 4759 scope.go:117] "RemoveContainer" containerID="224a9162353bccc985997aada1eb64d80abe62e27bdd63a9bc53ba850cb70353" Nov 25 19:46:31 crc kubenswrapper[4759]: I1125 19:46:31.059851 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vz6pj_67e7b861-9f72-41ba-ab98-35e0a2bdbe39/kube-multus/2.log" Nov 25 19:46:33 crc kubenswrapper[4759]: I1125 19:46:33.654348 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-84b47d7745-jnp5n" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.153349 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-747d477645-tgk7n" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.832078 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-989kk"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.833078 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.836539 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4h8cg" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.837341 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.841803 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-8tdxq"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.845440 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-989kk"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.845635 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.848802 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.848976 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.926486 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bcmbj"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.927509 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bcmbj" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.930387 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.930558 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.931689 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-njs5h" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.931961 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.942979 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-76794"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.944704 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.947996 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.948968 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-76794"] Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.996949 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m67d5\" (UniqueName: \"kubernetes.io/projected/95d5dc45-33bb-4278-a426-7534af845b52-kube-api-access-m67d5\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.996992 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-sockets\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997013 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5dc45-33bb-4278-a426-7534af845b52-cert\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997566 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-startup\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997609 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-reloader\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997642 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997665 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics-certs\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997721 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-conf\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:53 crc kubenswrapper[4759]: I1125 19:46:53.997784 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbqsw\" (UniqueName: \"kubernetes.io/projected/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-kube-api-access-hbqsw\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099264 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-metrics-certs\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099318 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-cert\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099351 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-sockets\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099379 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5dc45-33bb-4278-a426-7534af845b52-cert\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099402 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099430 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-startup\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099471 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-reloader\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099500 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099525 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics-certs\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099548 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metrics-certs\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099581 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-conf\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099728 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbqsw\" (UniqueName: \"kubernetes.io/projected/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-kube-api-access-hbqsw\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099792 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-985nz\" (UniqueName: \"kubernetes.io/projected/bc126569-65d5-4b4e-80cd-5b073b4faca9-kube-api-access-985nz\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099858 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jgwq\" (UniqueName: \"kubernetes.io/projected/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-kube-api-access-7jgwq\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099883 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metallb-excludel2\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099920 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-sockets\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.099936 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m67d5\" (UniqueName: \"kubernetes.io/projected/95d5dc45-33bb-4278-a426-7534af845b52-kube-api-access-m67d5\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.100002 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-conf\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.100162 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.100277 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-reloader\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.100551 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-frr-startup\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.105759 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5dc45-33bb-4278-a426-7534af845b52-cert\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.109091 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-metrics-certs\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.117546 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m67d5\" (UniqueName: \"kubernetes.io/projected/95d5dc45-33bb-4278-a426-7534af845b52-kube-api-access-m67d5\") pod \"frr-k8s-webhook-server-6998585d5-989kk\" (UID: \"95d5dc45-33bb-4278-a426-7534af845b52\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.120078 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbqsw\" (UniqueName: \"kubernetes.io/projected/f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a-kube-api-access-hbqsw\") pod \"frr-k8s-8tdxq\" (UID: \"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a\") " pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.153120 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.163082 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200335 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metrics-certs\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200398 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-985nz\" (UniqueName: \"kubernetes.io/projected/bc126569-65d5-4b4e-80cd-5b073b4faca9-kube-api-access-985nz\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200429 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jgwq\" (UniqueName: \"kubernetes.io/projected/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-kube-api-access-7jgwq\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200458 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metallb-excludel2\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200483 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-metrics-certs\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200501 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-cert\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.200524 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: E1125 19:46:54.200672 4759 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 19:46:54 crc kubenswrapper[4759]: E1125 19:46:54.200718 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist podName:d395fdc1-d2ad-413a-8cf9-8e6cf31e324e nodeName:}" failed. No retries permitted until 2025-11-25 19:46:54.700701709 +0000 UTC m=+805.286649410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist") pod "speaker-bcmbj" (UID: "d395fdc1-d2ad-413a-8cf9-8e6cf31e324e") : secret "metallb-memberlist" not found Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.201412 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metallb-excludel2\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.206296 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-metrics-certs\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.206437 4759 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.212931 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-metrics-certs\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.213999 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc126569-65d5-4b4e-80cd-5b073b4faca9-cert\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.220594 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jgwq\" (UniqueName: \"kubernetes.io/projected/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-kube-api-access-7jgwq\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.221898 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-985nz\" (UniqueName: \"kubernetes.io/projected/bc126569-65d5-4b4e-80cd-5b073b4faca9-kube-api-access-985nz\") pod \"controller-6c7b4b5f48-76794\" (UID: \"bc126569-65d5-4b4e-80cd-5b073b4faca9\") " pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.256185 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.342884 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-989kk"] Nov 25 19:46:54 crc kubenswrapper[4759]: W1125 19:46:54.350125 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95d5dc45_33bb_4278_a426_7534af845b52.slice/crio-6b6e4edae59b91b3b0ce0a9aaab1eaff52b1fe7be0aad2c3f6771a107224b7df WatchSource:0}: Error finding container 6b6e4edae59b91b3b0ce0a9aaab1eaff52b1fe7be0aad2c3f6771a107224b7df: Status 404 returned error can't find the container with id 6b6e4edae59b91b3b0ce0a9aaab1eaff52b1fe7be0aad2c3f6771a107224b7df Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.413667 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-76794"] Nov 25 19:46:54 crc kubenswrapper[4759]: W1125 19:46:54.416837 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc126569_65d5_4b4e_80cd_5b073b4faca9.slice/crio-8150f6258728b7dfa27b73c6b413c7743f8d9518fc41b48ec23edddd7dd898ec WatchSource:0}: Error finding container 8150f6258728b7dfa27b73c6b413c7743f8d9518fc41b48ec23edddd7dd898ec: Status 404 returned error can't find the container with id 8150f6258728b7dfa27b73c6b413c7743f8d9518fc41b48ec23edddd7dd898ec Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.705820 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.712120 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d395fdc1-d2ad-413a-8cf9-8e6cf31e324e-memberlist\") pod \"speaker-bcmbj\" (UID: \"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e\") " pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.845213 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bcmbj" Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.895454 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:46:54 crc kubenswrapper[4759]: I1125 19:46:54.895520 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.204056 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-76794" event={"ID":"bc126569-65d5-4b4e-80cd-5b073b4faca9","Type":"ContainerStarted","Data":"643c57ee35087757cc788c621a6a8d3933a5d0ee20cd76e61446713db8b47899"} Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.204116 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-76794" event={"ID":"bc126569-65d5-4b4e-80cd-5b073b4faca9","Type":"ContainerStarted","Data":"8150f6258728b7dfa27b73c6b413c7743f8d9518fc41b48ec23edddd7dd898ec"} Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.205155 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" event={"ID":"95d5dc45-33bb-4278-a426-7534af845b52","Type":"ContainerStarted","Data":"6b6e4edae59b91b3b0ce0a9aaab1eaff52b1fe7be0aad2c3f6771a107224b7df"} Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.206197 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bcmbj" event={"ID":"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e","Type":"ContainerStarted","Data":"0e7445da0e580984dba59835f2cbe79aca38d6b7a52b5fa651fc1c3db64e7121"} Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.206229 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bcmbj" event={"ID":"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e","Type":"ContainerStarted","Data":"0704369a964c8c65baeb19adadf8e40637cff50ee91fe37626a5be65384244f9"} Nov 25 19:46:55 crc kubenswrapper[4759]: I1125 19:46:55.206946 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"4fd7d648720bae966cd4e5e3d773d2091cefdb5fcf766706988a16594eeed8b3"} Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.229336 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-76794" event={"ID":"bc126569-65d5-4b4e-80cd-5b073b4faca9","Type":"ContainerStarted","Data":"482854489c5b732029bd196d713980ad9ab862f2aad683808495d295eff1be6b"} Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.229904 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.234353 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bcmbj" event={"ID":"d395fdc1-d2ad-413a-8cf9-8e6cf31e324e","Type":"ContainerStarted","Data":"96a06ee2a3246743ff1302a7f4908d5e2ad2dac6c6d53997749e773a5c3417d3"} Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.234551 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bcmbj" Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.249976 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-76794" podStartSLOduration=2.47624529 podStartE2EDuration="4.249953569s" podCreationTimestamp="2025-11-25 19:46:53 +0000 UTC" firstStartedPulling="2025-11-25 19:46:54.501261817 +0000 UTC m=+805.087209508" lastFinishedPulling="2025-11-25 19:46:56.274970086 +0000 UTC m=+806.860917787" observedRunningTime="2025-11-25 19:46:57.24749516 +0000 UTC m=+807.833442861" watchObservedRunningTime="2025-11-25 19:46:57.249953569 +0000 UTC m=+807.835901280" Nov 25 19:46:57 crc kubenswrapper[4759]: I1125 19:46:57.324065 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bcmbj" podStartSLOduration=3.134357286 podStartE2EDuration="4.324040608s" podCreationTimestamp="2025-11-25 19:46:53 +0000 UTC" firstStartedPulling="2025-11-25 19:46:55.087520027 +0000 UTC m=+805.673467728" lastFinishedPulling="2025-11-25 19:46:56.277203349 +0000 UTC m=+806.863151050" observedRunningTime="2025-11-25 19:46:57.322217857 +0000 UTC m=+807.908165558" watchObservedRunningTime="2025-11-25 19:46:57.324040608 +0000 UTC m=+807.909988309" Nov 25 19:47:02 crc kubenswrapper[4759]: I1125 19:47:02.278223 4759 generic.go:334] "Generic (PLEG): container finished" podID="f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a" containerID="d48fed041cd218ea9ce8a355b9139dd5b4cdcd869c4b8fb9f172ccf08e443081" exitCode=0 Nov 25 19:47:02 crc kubenswrapper[4759]: I1125 19:47:02.278314 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerDied","Data":"d48fed041cd218ea9ce8a355b9139dd5b4cdcd869c4b8fb9f172ccf08e443081"} Nov 25 19:47:02 crc kubenswrapper[4759]: I1125 19:47:02.281293 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" event={"ID":"95d5dc45-33bb-4278-a426-7534af845b52","Type":"ContainerStarted","Data":"22465c6352dd253596ba1b4da34fba6b9ab8ae5d460edc06195fa82db4816d48"} Nov 25 19:47:02 crc kubenswrapper[4759]: I1125 19:47:02.281570 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:47:02 crc kubenswrapper[4759]: I1125 19:47:02.318037 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" podStartSLOduration=2.249209587 podStartE2EDuration="9.318017489s" podCreationTimestamp="2025-11-25 19:46:53 +0000 UTC" firstStartedPulling="2025-11-25 19:46:54.352275745 +0000 UTC m=+804.938223446" lastFinishedPulling="2025-11-25 19:47:01.421083607 +0000 UTC m=+812.007031348" observedRunningTime="2025-11-25 19:47:02.312526675 +0000 UTC m=+812.898474376" watchObservedRunningTime="2025-11-25 19:47:02.318017489 +0000 UTC m=+812.903965190" Nov 25 19:47:03 crc kubenswrapper[4759]: I1125 19:47:03.288599 4759 generic.go:334] "Generic (PLEG): container finished" podID="f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a" containerID="87066680f8f1667b20c96b5515f5c48b2ff4d9991ccf8521e4df25ceb7b52514" exitCode=0 Nov 25 19:47:03 crc kubenswrapper[4759]: I1125 19:47:03.288756 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerDied","Data":"87066680f8f1667b20c96b5515f5c48b2ff4d9991ccf8521e4df25ceb7b52514"} Nov 25 19:47:04 crc kubenswrapper[4759]: I1125 19:47:04.259861 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-76794" Nov 25 19:47:04 crc kubenswrapper[4759]: I1125 19:47:04.297027 4759 generic.go:334] "Generic (PLEG): container finished" podID="f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a" containerID="3624bc281dcf614569c1d72b5b2bf30280df89c2ac87aabf54ee16eee81687be" exitCode=0 Nov 25 19:47:04 crc kubenswrapper[4759]: I1125 19:47:04.297076 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerDied","Data":"3624bc281dcf614569c1d72b5b2bf30280df89c2ac87aabf54ee16eee81687be"} Nov 25 19:47:05 crc kubenswrapper[4759]: I1125 19:47:05.306626 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"c5e1afc3b80b8efdebdcf818fc9c3e1ae4891fb1449428b7dd16206e678cb83c"} Nov 25 19:47:05 crc kubenswrapper[4759]: I1125 19:47:05.306943 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"11a9af33c8d649b1a021dfc713d2d8183a363564b95f90beb3c528bfca53b191"} Nov 25 19:47:05 crc kubenswrapper[4759]: I1125 19:47:05.306954 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"40b1441d5cf050ac0646988f59db35ca27691653ab3b8366ecc22c86cd353e54"} Nov 25 19:47:05 crc kubenswrapper[4759]: I1125 19:47:05.306962 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"7823d64a98b47075e5bad4a7d5eb051b84ed5423ac3200b7173b8dac565a3987"} Nov 25 19:47:05 crc kubenswrapper[4759]: I1125 19:47:05.306970 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"791f733b8d0093d1a1b43bd43b5e95b8b6e9d285d2b17e35b6461c5cfeb40066"} Nov 25 19:47:06 crc kubenswrapper[4759]: I1125 19:47:06.329790 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8tdxq" event={"ID":"f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a","Type":"ContainerStarted","Data":"030135457d115da0bb18caafb90f852d56473d03c12b6f279d3411517cb18ba5"} Nov 25 19:47:06 crc kubenswrapper[4759]: I1125 19:47:06.330274 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:47:06 crc kubenswrapper[4759]: I1125 19:47:06.358899 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-8tdxq" podStartSLOduration=6.206853631 podStartE2EDuration="13.3588802s" podCreationTimestamp="2025-11-25 19:46:53 +0000 UTC" firstStartedPulling="2025-11-25 19:46:54.29691315 +0000 UTC m=+804.882860851" lastFinishedPulling="2025-11-25 19:47:01.448939719 +0000 UTC m=+812.034887420" observedRunningTime="2025-11-25 19:47:06.35534846 +0000 UTC m=+816.941296161" watchObservedRunningTime="2025-11-25 19:47:06.3588802 +0000 UTC m=+816.944827901" Nov 25 19:47:09 crc kubenswrapper[4759]: I1125 19:47:09.163964 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:47:09 crc kubenswrapper[4759]: I1125 19:47:09.212259 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:47:14 crc kubenswrapper[4759]: I1125 19:47:14.160658 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-989kk" Nov 25 19:47:14 crc kubenswrapper[4759]: I1125 19:47:14.164908 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-8tdxq" Nov 25 19:47:14 crc kubenswrapper[4759]: I1125 19:47:14.848540 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bcmbj" Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.771660 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.772578 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.774187 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.774967 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.777158 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-4wtt2" Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.782983 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:20 crc kubenswrapper[4759]: I1125 19:47:20.949234 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsgbc\" (UniqueName: \"kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc\") pod \"mariadb-operator-index-b6lb9\" (UID: \"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b\") " pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:21 crc kubenswrapper[4759]: I1125 19:47:21.050090 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgbc\" (UniqueName: \"kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc\") pod \"mariadb-operator-index-b6lb9\" (UID: \"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b\") " pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:21 crc kubenswrapper[4759]: I1125 19:47:21.072243 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgbc\" (UniqueName: \"kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc\") pod \"mariadb-operator-index-b6lb9\" (UID: \"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b\") " pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:21 crc kubenswrapper[4759]: I1125 19:47:21.090103 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:21 crc kubenswrapper[4759]: I1125 19:47:21.325375 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:21 crc kubenswrapper[4759]: I1125 19:47:21.413110 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-b6lb9" event={"ID":"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b","Type":"ContainerStarted","Data":"ba120631a53310d0560c5944bf61a778dafb967a14f38dac4952a4688c44e1ed"} Nov 25 19:47:23 crc kubenswrapper[4759]: I1125 19:47:23.430956 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-b6lb9" event={"ID":"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b","Type":"ContainerStarted","Data":"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2"} Nov 25 19:47:23 crc kubenswrapper[4759]: I1125 19:47:23.447792 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-b6lb9" podStartSLOduration=2.335813797 podStartE2EDuration="3.447775236s" podCreationTimestamp="2025-11-25 19:47:20 +0000 UTC" firstStartedPulling="2025-11-25 19:47:21.333175957 +0000 UTC m=+831.919123658" lastFinishedPulling="2025-11-25 19:47:22.445137386 +0000 UTC m=+833.031085097" observedRunningTime="2025-11-25 19:47:23.444639588 +0000 UTC m=+834.030587289" watchObservedRunningTime="2025-11-25 19:47:23.447775236 +0000 UTC m=+834.033722937" Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.152571 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.766066 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-whf2r"] Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.767719 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.770060 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-whf2r"] Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.895574 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.895629 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.895670 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.896208 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.896265 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18" gracePeriod=600 Nov 25 19:47:24 crc kubenswrapper[4759]: I1125 19:47:24.902541 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlmvf\" (UniqueName: \"kubernetes.io/projected/d3320894-675c-4b39-a183-251eda3f4cce-kube-api-access-nlmvf\") pod \"mariadb-operator-index-whf2r\" (UID: \"d3320894-675c-4b39-a183-251eda3f4cce\") " pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.003979 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlmvf\" (UniqueName: \"kubernetes.io/projected/d3320894-675c-4b39-a183-251eda3f4cce-kube-api-access-nlmvf\") pod \"mariadb-operator-index-whf2r\" (UID: \"d3320894-675c-4b39-a183-251eda3f4cce\") " pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.024107 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlmvf\" (UniqueName: \"kubernetes.io/projected/d3320894-675c-4b39-a183-251eda3f4cce-kube-api-access-nlmvf\") pod \"mariadb-operator-index-whf2r\" (UID: \"d3320894-675c-4b39-a183-251eda3f4cce\") " pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.092465 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.262803 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-whf2r"] Nov 25 19:47:25 crc kubenswrapper[4759]: W1125 19:47:25.269391 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3320894_675c_4b39_a183_251eda3f4cce.slice/crio-fe1939539e687b6639140910d0d07d39d1aef33c11d0595db1a7ccb1a2d9261a WatchSource:0}: Error finding container fe1939539e687b6639140910d0d07d39d1aef33c11d0595db1a7ccb1a2d9261a: Status 404 returned error can't find the container with id fe1939539e687b6639140910d0d07d39d1aef33c11d0595db1a7ccb1a2d9261a Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.441633 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-whf2r" event={"ID":"d3320894-675c-4b39-a183-251eda3f4cce","Type":"ContainerStarted","Data":"fe1939539e687b6639140910d0d07d39d1aef33c11d0595db1a7ccb1a2d9261a"} Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.444405 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18" exitCode=0 Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.444496 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18"} Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.444549 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8"} Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.444569 4759 scope.go:117] "RemoveContainer" containerID="3d2365e9a70cd4d56387270569c148782191d916315817fcee7b66d279d72a1b" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.444569 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-b6lb9" podUID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" containerName="registry-server" containerID="cri-o://d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2" gracePeriod=2 Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.750148 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.913846 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsgbc\" (UniqueName: \"kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc\") pod \"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b\" (UID: \"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b\") " Nov 25 19:47:25 crc kubenswrapper[4759]: I1125 19:47:25.918684 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc" (OuterVolumeSpecName: "kube-api-access-dsgbc") pod "fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" (UID: "fcc88fbe-5ca3-42d6-b7d1-4163748ce11b"). InnerVolumeSpecName "kube-api-access-dsgbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.015176 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsgbc\" (UniqueName: \"kubernetes.io/projected/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b-kube-api-access-dsgbc\") on node \"crc\" DevicePath \"\"" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.453738 4759 generic.go:334] "Generic (PLEG): container finished" podID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" containerID="d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2" exitCode=0 Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.453811 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-b6lb9" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.453826 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-b6lb9" event={"ID":"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b","Type":"ContainerDied","Data":"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2"} Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.454281 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-b6lb9" event={"ID":"fcc88fbe-5ca3-42d6-b7d1-4163748ce11b","Type":"ContainerDied","Data":"ba120631a53310d0560c5944bf61a778dafb967a14f38dac4952a4688c44e1ed"} Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.454317 4759 scope.go:117] "RemoveContainer" containerID="d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.456886 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-whf2r" event={"ID":"d3320894-675c-4b39-a183-251eda3f4cce","Type":"ContainerStarted","Data":"63bf2895cfe40cfbca49dd1820f4ef4ed8337c0817e98e920605149bf903a247"} Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.471666 4759 scope.go:117] "RemoveContainer" containerID="d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2" Nov 25 19:47:26 crc kubenswrapper[4759]: E1125 19:47:26.475155 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2\": container with ID starting with d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2 not found: ID does not exist" containerID="d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.475203 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2"} err="failed to get container status \"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2\": rpc error: code = NotFound desc = could not find container \"d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2\": container with ID starting with d0ae39ff1ac3da6d4099963c293504a18d912155289adebe12da9df08e87acd2 not found: ID does not exist" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.477778 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-whf2r" podStartSLOduration=2.086359976 podStartE2EDuration="2.477758244s" podCreationTimestamp="2025-11-25 19:47:24 +0000 UTC" firstStartedPulling="2025-11-25 19:47:25.273934737 +0000 UTC m=+835.859882438" lastFinishedPulling="2025-11-25 19:47:25.665333005 +0000 UTC m=+836.251280706" observedRunningTime="2025-11-25 19:47:26.474249316 +0000 UTC m=+837.060197017" watchObservedRunningTime="2025-11-25 19:47:26.477758244 +0000 UTC m=+837.063705945" Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.497855 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:26 crc kubenswrapper[4759]: I1125 19:47:26.506021 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-b6lb9"] Nov 25 19:47:28 crc kubenswrapper[4759]: I1125 19:47:28.114086 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" path="/var/lib/kubelet/pods/fcc88fbe-5ca3-42d6-b7d1-4163748ce11b/volumes" Nov 25 19:47:35 crc kubenswrapper[4759]: I1125 19:47:35.093202 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:35 crc kubenswrapper[4759]: I1125 19:47:35.093720 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:35 crc kubenswrapper[4759]: I1125 19:47:35.118179 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:35 crc kubenswrapper[4759]: I1125 19:47:35.577231 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-whf2r" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.218850 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w"] Nov 25 19:47:40 crc kubenswrapper[4759]: E1125 19:47:40.219345 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" containerName="registry-server" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.219361 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" containerName="registry-server" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.219861 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc88fbe-5ca3-42d6-b7d1-4163748ce11b" containerName="registry-server" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.220626 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.222742 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.237233 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w"] Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.330115 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z966g\" (UniqueName: \"kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.330208 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.330260 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.432047 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z966g\" (UniqueName: \"kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.432097 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.432117 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.432604 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.432771 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.451773 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z966g\" (UniqueName: \"kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.543411 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:40 crc kubenswrapper[4759]: I1125 19:47:40.735874 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w"] Nov 25 19:47:41 crc kubenswrapper[4759]: I1125 19:47:41.556260 4759 generic.go:334] "Generic (PLEG): container finished" podID="11846abc-749c-4559-8f1b-38f0df47237b" containerID="5b7e16eca24b23b72c1bea821c1faf03bbfbdaa9884ff5fca73621385e5dcbf2" exitCode=0 Nov 25 19:47:41 crc kubenswrapper[4759]: I1125 19:47:41.556554 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" event={"ID":"11846abc-749c-4559-8f1b-38f0df47237b","Type":"ContainerDied","Data":"5b7e16eca24b23b72c1bea821c1faf03bbfbdaa9884ff5fca73621385e5dcbf2"} Nov 25 19:47:41 crc kubenswrapper[4759]: I1125 19:47:41.556586 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" event={"ID":"11846abc-749c-4559-8f1b-38f0df47237b","Type":"ContainerStarted","Data":"81315115bd572e6536a5bc43b9195c4759e33dba8882f1ef660b6a17c72714ce"} Nov 25 19:47:42 crc kubenswrapper[4759]: I1125 19:47:42.564188 4759 generic.go:334] "Generic (PLEG): container finished" podID="11846abc-749c-4559-8f1b-38f0df47237b" containerID="564fe891755706c397bc7e7e63cd0149c43cfd613fd9cccca95dccac477cfeb6" exitCode=0 Nov 25 19:47:42 crc kubenswrapper[4759]: I1125 19:47:42.564299 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" event={"ID":"11846abc-749c-4559-8f1b-38f0df47237b","Type":"ContainerDied","Data":"564fe891755706c397bc7e7e63cd0149c43cfd613fd9cccca95dccac477cfeb6"} Nov 25 19:47:43 crc kubenswrapper[4759]: I1125 19:47:43.578846 4759 generic.go:334] "Generic (PLEG): container finished" podID="11846abc-749c-4559-8f1b-38f0df47237b" containerID="e63225a95295a1547b1abf6337eb8558fd221572f70f9d81f72d26885913656b" exitCode=0 Nov 25 19:47:43 crc kubenswrapper[4759]: I1125 19:47:43.578967 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" event={"ID":"11846abc-749c-4559-8f1b-38f0df47237b","Type":"ContainerDied","Data":"e63225a95295a1547b1abf6337eb8558fd221572f70f9d81f72d26885913656b"} Nov 25 19:47:44 crc kubenswrapper[4759]: I1125 19:47:44.904090 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.010994 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util\") pod \"11846abc-749c-4559-8f1b-38f0df47237b\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.011057 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z966g\" (UniqueName: \"kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g\") pod \"11846abc-749c-4559-8f1b-38f0df47237b\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.011087 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle\") pod \"11846abc-749c-4559-8f1b-38f0df47237b\" (UID: \"11846abc-749c-4559-8f1b-38f0df47237b\") " Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.012274 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle" (OuterVolumeSpecName: "bundle") pod "11846abc-749c-4559-8f1b-38f0df47237b" (UID: "11846abc-749c-4559-8f1b-38f0df47237b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.016360 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g" (OuterVolumeSpecName: "kube-api-access-z966g") pod "11846abc-749c-4559-8f1b-38f0df47237b" (UID: "11846abc-749c-4559-8f1b-38f0df47237b"). InnerVolumeSpecName "kube-api-access-z966g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.025775 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util" (OuterVolumeSpecName: "util") pod "11846abc-749c-4559-8f1b-38f0df47237b" (UID: "11846abc-749c-4559-8f1b-38f0df47237b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.112512 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.112546 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z966g\" (UniqueName: \"kubernetes.io/projected/11846abc-749c-4559-8f1b-38f0df47237b-kube-api-access-z966g\") on node \"crc\" DevicePath \"\"" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.112559 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11846abc-749c-4559-8f1b-38f0df47237b-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.595573 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" event={"ID":"11846abc-749c-4559-8f1b-38f0df47237b","Type":"ContainerDied","Data":"81315115bd572e6536a5bc43b9195c4759e33dba8882f1ef660b6a17c72714ce"} Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.595615 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81315115bd572e6536a5bc43b9195c4759e33dba8882f1ef660b6a17c72714ce" Nov 25 19:47:45 crc kubenswrapper[4759]: I1125 19:47:45.595727 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.232002 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8"] Nov 25 19:47:49 crc kubenswrapper[4759]: E1125 19:47:49.232559 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="pull" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.232577 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="pull" Nov 25 19:47:49 crc kubenswrapper[4759]: E1125 19:47:49.232599 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="util" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.232606 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="util" Nov 25 19:47:49 crc kubenswrapper[4759]: E1125 19:47:49.232616 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="extract" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.232624 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="extract" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.232735 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="11846abc-749c-4559-8f1b-38f0df47237b" containerName="extract" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.233171 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.235007 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.235262 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.235395 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7nm7s" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.246714 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8"] Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.266202 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmlr4\" (UniqueName: \"kubernetes.io/projected/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-kube-api-access-qmlr4\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.266255 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-webhook-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.266309 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-apiservice-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.367034 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-apiservice-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.367152 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-webhook-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.367183 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmlr4\" (UniqueName: \"kubernetes.io/projected/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-kube-api-access-qmlr4\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.372525 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-apiservice-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.376164 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-webhook-cert\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.382796 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmlr4\" (UniqueName: \"kubernetes.io/projected/4afc0c17-9efa-4f7c-bd72-c2a3091eeddd-kube-api-access-qmlr4\") pod \"mariadb-operator-controller-manager-847fcc49c9-t9qj8\" (UID: \"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd\") " pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.553980 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:49 crc kubenswrapper[4759]: I1125 19:47:49.959663 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8"] Nov 25 19:47:50 crc kubenswrapper[4759]: I1125 19:47:50.622677 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" event={"ID":"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd","Type":"ContainerStarted","Data":"a89925ea7fe477a81368e433b3726e12bef990a2f12b7930288a7c6e310420b9"} Nov 25 19:47:53 crc kubenswrapper[4759]: I1125 19:47:53.639757 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" event={"ID":"4afc0c17-9efa-4f7c-bd72-c2a3091eeddd","Type":"ContainerStarted","Data":"c936009cc1b4cd7a6e463fa69424cbe1c3ad3815e76d5f7eaaba4b8c28dd8fbd"} Nov 25 19:47:53 crc kubenswrapper[4759]: I1125 19:47:53.640348 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:47:53 crc kubenswrapper[4759]: I1125 19:47:53.659098 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" podStartSLOduration=1.208050701 podStartE2EDuration="4.659077817s" podCreationTimestamp="2025-11-25 19:47:49 +0000 UTC" firstStartedPulling="2025-11-25 19:47:49.967038494 +0000 UTC m=+860.552986195" lastFinishedPulling="2025-11-25 19:47:53.41806561 +0000 UTC m=+864.004013311" observedRunningTime="2025-11-25 19:47:53.655778838 +0000 UTC m=+864.241726559" watchObservedRunningTime="2025-11-25 19:47:53.659077817 +0000 UTC m=+864.245025518" Nov 25 19:47:59 crc kubenswrapper[4759]: I1125 19:47:59.558980 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-847fcc49c9-t9qj8" Nov 25 19:48:05 crc kubenswrapper[4759]: I1125 19:48:05.824477 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:05 crc kubenswrapper[4759]: I1125 19:48:05.826130 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:05 crc kubenswrapper[4759]: I1125 19:48:05.829256 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-5dbdn" Nov 25 19:48:05 crc kubenswrapper[4759]: I1125 19:48:05.847778 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.013067 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwqdv\" (UniqueName: \"kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv\") pod \"infra-operator-index-n47vj\" (UID: \"f908b5d0-404f-4a28-b4cc-30048c035b71\") " pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.113823 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwqdv\" (UniqueName: \"kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv\") pod \"infra-operator-index-n47vj\" (UID: \"f908b5d0-404f-4a28-b4cc-30048c035b71\") " pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.136134 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwqdv\" (UniqueName: \"kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv\") pod \"infra-operator-index-n47vj\" (UID: \"f908b5d0-404f-4a28-b4cc-30048c035b71\") " pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.145903 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.556035 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:06 crc kubenswrapper[4759]: W1125 19:48:06.562633 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf908b5d0_404f_4a28_b4cc_30048c035b71.slice/crio-1a41313a996a1d1d1ef17c8756261cc3bd52601c2f8234ffa40881766b40d794 WatchSource:0}: Error finding container 1a41313a996a1d1d1ef17c8756261cc3bd52601c2f8234ffa40881766b40d794: Status 404 returned error can't find the container with id 1a41313a996a1d1d1ef17c8756261cc3bd52601c2f8234ffa40881766b40d794 Nov 25 19:48:06 crc kubenswrapper[4759]: I1125 19:48:06.726920 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-n47vj" event={"ID":"f908b5d0-404f-4a28-b4cc-30048c035b71","Type":"ContainerStarted","Data":"1a41313a996a1d1d1ef17c8756261cc3bd52601c2f8234ffa40881766b40d794"} Nov 25 19:48:07 crc kubenswrapper[4759]: I1125 19:48:07.735203 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-n47vj" event={"ID":"f908b5d0-404f-4a28-b4cc-30048c035b71","Type":"ContainerStarted","Data":"eef3f5c396a1d5003c62cc2c9fc423260b461bc982ca97f1a43bc33318577d32"} Nov 25 19:48:07 crc kubenswrapper[4759]: I1125 19:48:07.755616 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-n47vj" podStartSLOduration=1.98230809 podStartE2EDuration="2.755589845s" podCreationTimestamp="2025-11-25 19:48:05 +0000 UTC" firstStartedPulling="2025-11-25 19:48:06.566611265 +0000 UTC m=+877.152558966" lastFinishedPulling="2025-11-25 19:48:07.33989302 +0000 UTC m=+877.925840721" observedRunningTime="2025-11-25 19:48:07.752235145 +0000 UTC m=+878.338182866" watchObservedRunningTime="2025-11-25 19:48:07.755589845 +0000 UTC m=+878.341537546" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.034165 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.036689 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.044492 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.150394 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.150520 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.150895 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46w6f\" (UniqueName: \"kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.252187 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46w6f\" (UniqueName: \"kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.252307 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.252361 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.253260 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.253358 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.286148 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46w6f\" (UniqueName: \"kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f\") pod \"certified-operators-lx7lg\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.357935 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.662057 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:08 crc kubenswrapper[4759]: I1125 19:48:08.741933 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerStarted","Data":"fad7196164ea8b2f3cc2342844041278445fda7a3a53b0f985d94a62bf30eeeb"} Nov 25 19:48:09 crc kubenswrapper[4759]: I1125 19:48:09.749108 4759 generic.go:334] "Generic (PLEG): container finished" podID="5149f920-1f79-4694-9e83-c693bb5036d1" containerID="7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654" exitCode=0 Nov 25 19:48:09 crc kubenswrapper[4759]: I1125 19:48:09.749226 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerDied","Data":"7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654"} Nov 25 19:48:10 crc kubenswrapper[4759]: I1125 19:48:10.756518 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerStarted","Data":"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d"} Nov 25 19:48:10 crc kubenswrapper[4759]: E1125 19:48:10.783477 4759 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5149f920_1f79_4694_9e83_c693bb5036d1.slice/crio-0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5149f920_1f79_4694_9e83_c693bb5036d1.slice/crio-conmon-0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d.scope\": RecentStats: unable to find data in memory cache]" Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.618892 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.619435 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-n47vj" podUID="f908b5d0-404f-4a28-b4cc-30048c035b71" containerName="registry-server" containerID="cri-o://eef3f5c396a1d5003c62cc2c9fc423260b461bc982ca97f1a43bc33318577d32" gracePeriod=2 Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.765162 4759 generic.go:334] "Generic (PLEG): container finished" podID="5149f920-1f79-4694-9e83-c693bb5036d1" containerID="0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d" exitCode=0 Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.765224 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerDied","Data":"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d"} Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.765268 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerStarted","Data":"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2"} Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.768861 4759 generic.go:334] "Generic (PLEG): container finished" podID="f908b5d0-404f-4a28-b4cc-30048c035b71" containerID="eef3f5c396a1d5003c62cc2c9fc423260b461bc982ca97f1a43bc33318577d32" exitCode=0 Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.768976 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-n47vj" event={"ID":"f908b5d0-404f-4a28-b4cc-30048c035b71","Type":"ContainerDied","Data":"eef3f5c396a1d5003c62cc2c9fc423260b461bc982ca97f1a43bc33318577d32"} Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.794696 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lx7lg" podStartSLOduration=2.385374874 podStartE2EDuration="3.794681373s" podCreationTimestamp="2025-11-25 19:48:08 +0000 UTC" firstStartedPulling="2025-11-25 19:48:09.751543489 +0000 UTC m=+880.337491210" lastFinishedPulling="2025-11-25 19:48:11.160850008 +0000 UTC m=+881.746797709" observedRunningTime="2025-11-25 19:48:11.791236861 +0000 UTC m=+882.377184562" watchObservedRunningTime="2025-11-25 19:48:11.794681373 +0000 UTC m=+882.380629074" Nov 25 19:48:11 crc kubenswrapper[4759]: I1125 19:48:11.983363 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.106967 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwqdv\" (UniqueName: \"kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv\") pod \"f908b5d0-404f-4a28-b4cc-30048c035b71\" (UID: \"f908b5d0-404f-4a28-b4cc-30048c035b71\") " Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.115044 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv" (OuterVolumeSpecName: "kube-api-access-dwqdv") pod "f908b5d0-404f-4a28-b4cc-30048c035b71" (UID: "f908b5d0-404f-4a28-b4cc-30048c035b71"). InnerVolumeSpecName "kube-api-access-dwqdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.208698 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwqdv\" (UniqueName: \"kubernetes.io/projected/f908b5d0-404f-4a28-b4cc-30048c035b71-kube-api-access-dwqdv\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.424471 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-5mwnr"] Nov 25 19:48:12 crc kubenswrapper[4759]: E1125 19:48:12.424687 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f908b5d0-404f-4a28-b4cc-30048c035b71" containerName="registry-server" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.424699 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="f908b5d0-404f-4a28-b4cc-30048c035b71" containerName="registry-server" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.424804 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="f908b5d0-404f-4a28-b4cc-30048c035b71" containerName="registry-server" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.425169 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.439994 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-5mwnr"] Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.613398 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkmnj\" (UniqueName: \"kubernetes.io/projected/224d5f03-e4d9-4652-bfc0-e7eb6bb452c1-kube-api-access-nkmnj\") pod \"infra-operator-index-5mwnr\" (UID: \"224d5f03-e4d9-4652-bfc0-e7eb6bb452c1\") " pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.714985 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkmnj\" (UniqueName: \"kubernetes.io/projected/224d5f03-e4d9-4652-bfc0-e7eb6bb452c1-kube-api-access-nkmnj\") pod \"infra-operator-index-5mwnr\" (UID: \"224d5f03-e4d9-4652-bfc0-e7eb6bb452c1\") " pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.730873 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkmnj\" (UniqueName: \"kubernetes.io/projected/224d5f03-e4d9-4652-bfc0-e7eb6bb452c1-kube-api-access-nkmnj\") pod \"infra-operator-index-5mwnr\" (UID: \"224d5f03-e4d9-4652-bfc0-e7eb6bb452c1\") " pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.753260 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.776095 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-n47vj" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.776128 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-n47vj" event={"ID":"f908b5d0-404f-4a28-b4cc-30048c035b71","Type":"ContainerDied","Data":"1a41313a996a1d1d1ef17c8756261cc3bd52601c2f8234ffa40881766b40d794"} Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.776214 4759 scope.go:117] "RemoveContainer" containerID="eef3f5c396a1d5003c62cc2c9fc423260b461bc982ca97f1a43bc33318577d32" Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.819113 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:12 crc kubenswrapper[4759]: I1125 19:48:12.824169 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-n47vj"] Nov 25 19:48:13 crc kubenswrapper[4759]: I1125 19:48:13.206297 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-5mwnr"] Nov 25 19:48:13 crc kubenswrapper[4759]: I1125 19:48:13.783773 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-5mwnr" event={"ID":"224d5f03-e4d9-4652-bfc0-e7eb6bb452c1","Type":"ContainerStarted","Data":"dd47d3fff3938752162338f7024cfef11096a9635c5ce65e834d5fd794b4479c"} Nov 25 19:48:14 crc kubenswrapper[4759]: I1125 19:48:14.118005 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f908b5d0-404f-4a28-b4cc-30048c035b71" path="/var/lib/kubelet/pods/f908b5d0-404f-4a28-b4cc-30048c035b71/volumes" Nov 25 19:48:14 crc kubenswrapper[4759]: I1125 19:48:14.792770 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-5mwnr" event={"ID":"224d5f03-e4d9-4652-bfc0-e7eb6bb452c1","Type":"ContainerStarted","Data":"1c0e8b831b1679bb34d807759d9fa9c03cde28d8373abad3fb0dc418afa10ec5"} Nov 25 19:48:14 crc kubenswrapper[4759]: I1125 19:48:14.810422 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-5mwnr" podStartSLOduration=2.32793301 podStartE2EDuration="2.810402616s" podCreationTimestamp="2025-11-25 19:48:12 +0000 UTC" firstStartedPulling="2025-11-25 19:48:13.228266097 +0000 UTC m=+883.814213798" lastFinishedPulling="2025-11-25 19:48:13.710735703 +0000 UTC m=+884.296683404" observedRunningTime="2025-11-25 19:48:14.806298115 +0000 UTC m=+885.392245846" watchObservedRunningTime="2025-11-25 19:48:14.810402616 +0000 UTC m=+885.396350317" Nov 25 19:48:18 crc kubenswrapper[4759]: I1125 19:48:18.358170 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:18 crc kubenswrapper[4759]: I1125 19:48:18.358510 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:18 crc kubenswrapper[4759]: I1125 19:48:18.428717 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:18 crc kubenswrapper[4759]: I1125 19:48:18.845814 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.017216 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.017770 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lx7lg" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="registry-server" containerID="cri-o://378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2" gracePeriod=2 Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.404224 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.464408 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46w6f\" (UniqueName: \"kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f\") pod \"5149f920-1f79-4694-9e83-c693bb5036d1\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.464500 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content\") pod \"5149f920-1f79-4694-9e83-c693bb5036d1\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.464548 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities\") pod \"5149f920-1f79-4694-9e83-c693bb5036d1\" (UID: \"5149f920-1f79-4694-9e83-c693bb5036d1\") " Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.465486 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities" (OuterVolumeSpecName: "utilities") pod "5149f920-1f79-4694-9e83-c693bb5036d1" (UID: "5149f920-1f79-4694-9e83-c693bb5036d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.469678 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f" (OuterVolumeSpecName: "kube-api-access-46w6f") pod "5149f920-1f79-4694-9e83-c693bb5036d1" (UID: "5149f920-1f79-4694-9e83-c693bb5036d1"). InnerVolumeSpecName "kube-api-access-46w6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.509712 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5149f920-1f79-4694-9e83-c693bb5036d1" (UID: "5149f920-1f79-4694-9e83-c693bb5036d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.566369 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46w6f\" (UniqueName: \"kubernetes.io/projected/5149f920-1f79-4694-9e83-c693bb5036d1-kube-api-access-46w6f\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.566404 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.566415 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149f920-1f79-4694-9e83-c693bb5036d1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.829786 4759 generic.go:334] "Generic (PLEG): container finished" podID="5149f920-1f79-4694-9e83-c693bb5036d1" containerID="378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2" exitCode=0 Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.829843 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerDied","Data":"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2"} Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.829897 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7lg" event={"ID":"5149f920-1f79-4694-9e83-c693bb5036d1","Type":"ContainerDied","Data":"fad7196164ea8b2f3cc2342844041278445fda7a3a53b0f985d94a62bf30eeeb"} Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.829920 4759 scope.go:117] "RemoveContainer" containerID="378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.829857 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7lg" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.850079 4759 scope.go:117] "RemoveContainer" containerID="0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.861520 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.866309 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lx7lg"] Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.888150 4759 scope.go:117] "RemoveContainer" containerID="7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.907536 4759 scope.go:117] "RemoveContainer" containerID="378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2" Nov 25 19:48:21 crc kubenswrapper[4759]: E1125 19:48:21.908036 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2\": container with ID starting with 378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2 not found: ID does not exist" containerID="378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.908082 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2"} err="failed to get container status \"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2\": rpc error: code = NotFound desc = could not find container \"378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2\": container with ID starting with 378c6c9f7cc14f85440e2d59c551e62d44d80005836e79f8a122c58fab8747a2 not found: ID does not exist" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.908115 4759 scope.go:117] "RemoveContainer" containerID="0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d" Nov 25 19:48:21 crc kubenswrapper[4759]: E1125 19:48:21.908487 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d\": container with ID starting with 0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d not found: ID does not exist" containerID="0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.908520 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d"} err="failed to get container status \"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d\": rpc error: code = NotFound desc = could not find container \"0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d\": container with ID starting with 0f915884ef3998e8bda6cef0c0b0144dfbc0d87f305723c8509311374074f03d not found: ID does not exist" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.908543 4759 scope.go:117] "RemoveContainer" containerID="7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654" Nov 25 19:48:21 crc kubenswrapper[4759]: E1125 19:48:21.908849 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654\": container with ID starting with 7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654 not found: ID does not exist" containerID="7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654" Nov 25 19:48:21 crc kubenswrapper[4759]: I1125 19:48:21.908877 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654"} err="failed to get container status \"7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654\": rpc error: code = NotFound desc = could not find container \"7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654\": container with ID starting with 7c82d3117711e2b09e786682f2d7f690a6900a7a142a87ce24aaad83bb520654 not found: ID does not exist" Nov 25 19:48:22 crc kubenswrapper[4759]: I1125 19:48:22.117779 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" path="/var/lib/kubelet/pods/5149f920-1f79-4694-9e83-c693bb5036d1/volumes" Nov 25 19:48:22 crc kubenswrapper[4759]: I1125 19:48:22.753871 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:22 crc kubenswrapper[4759]: I1125 19:48:22.753936 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:22 crc kubenswrapper[4759]: I1125 19:48:22.805415 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:22 crc kubenswrapper[4759]: I1125 19:48:22.869537 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-5mwnr" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.871664 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp"] Nov 25 19:48:26 crc kubenswrapper[4759]: E1125 19:48:26.872330 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="extract-utilities" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.872350 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="extract-utilities" Nov 25 19:48:26 crc kubenswrapper[4759]: E1125 19:48:26.872381 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="registry-server" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.872392 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="registry-server" Nov 25 19:48:26 crc kubenswrapper[4759]: E1125 19:48:26.872415 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="extract-content" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.872425 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="extract-content" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.872612 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="5149f920-1f79-4694-9e83-c693bb5036d1" containerName="registry-server" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.873719 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.876311 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:48:26 crc kubenswrapper[4759]: I1125 19:48:26.888894 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp"] Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.029141 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llhsf\" (UniqueName: \"kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.029371 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.029503 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.131289 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llhsf\" (UniqueName: \"kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.131358 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.131495 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.132069 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.132154 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.161744 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llhsf\" (UniqueName: \"kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.201889 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.638099 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp"] Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.874011 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerStarted","Data":"d782e71979673b632636c31526a0fa054eeef7791ada13ed276acf1228207c62"} Nov 25 19:48:27 crc kubenswrapper[4759]: I1125 19:48:27.875289 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerStarted","Data":"7f664d9db99e7ccf5895fc78ee705e270a2785d976b1d66d4e3e90835b4bcec1"} Nov 25 19:48:28 crc kubenswrapper[4759]: I1125 19:48:28.883565 4759 generic.go:334] "Generic (PLEG): container finished" podID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerID="d782e71979673b632636c31526a0fa054eeef7791ada13ed276acf1228207c62" exitCode=0 Nov 25 19:48:28 crc kubenswrapper[4759]: I1125 19:48:28.883606 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerDied","Data":"d782e71979673b632636c31526a0fa054eeef7791ada13ed276acf1228207c62"} Nov 25 19:48:29 crc kubenswrapper[4759]: I1125 19:48:29.893367 4759 generic.go:334] "Generic (PLEG): container finished" podID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerID="b1087d5fb8e6d0e283750a16140447d59ca450b9428745bf0c9285685c2d1660" exitCode=0 Nov 25 19:48:29 crc kubenswrapper[4759]: I1125 19:48:29.893488 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerDied","Data":"b1087d5fb8e6d0e283750a16140447d59ca450b9428745bf0c9285685c2d1660"} Nov 25 19:48:30 crc kubenswrapper[4759]: I1125 19:48:30.901340 4759 generic.go:334] "Generic (PLEG): container finished" podID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerID="28592e2fe14f5e7173b9de8b7dc9bcf9fc933daa5d47c7eb11eb13341d72883c" exitCode=0 Nov 25 19:48:30 crc kubenswrapper[4759]: I1125 19:48:30.901383 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerDied","Data":"28592e2fe14f5e7173b9de8b7dc9bcf9fc933daa5d47c7eb11eb13341d72883c"} Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.163467 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.297587 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llhsf\" (UniqueName: \"kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf\") pod \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.297851 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle\") pod \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.297997 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util\") pod \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\" (UID: \"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c\") " Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.298714 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle" (OuterVolumeSpecName: "bundle") pod "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" (UID: "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.303599 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf" (OuterVolumeSpecName: "kube-api-access-llhsf") pod "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" (UID: "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c"). InnerVolumeSpecName "kube-api-access-llhsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.318612 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util" (OuterVolumeSpecName: "util") pod "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" (UID: "b11f3e19-ffc6-4a83-83b1-25467cbcbc4c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.399925 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.399962 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llhsf\" (UniqueName: \"kubernetes.io/projected/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-kube-api-access-llhsf\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.399972 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b11f3e19-ffc6-4a83-83b1-25467cbcbc4c-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.916063 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" event={"ID":"b11f3e19-ffc6-4a83-83b1-25467cbcbc4c","Type":"ContainerDied","Data":"7f664d9db99e7ccf5895fc78ee705e270a2785d976b1d66d4e3e90835b4bcec1"} Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.916376 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f664d9db99e7ccf5895fc78ee705e270a2785d976b1d66d4e3e90835b4bcec1" Nov 25 19:48:32 crc kubenswrapper[4759]: I1125 19:48:32.916292 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.709083 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc"] Nov 25 19:48:39 crc kubenswrapper[4759]: E1125 19:48:39.709609 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="pull" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.709624 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="pull" Nov 25 19:48:39 crc kubenswrapper[4759]: E1125 19:48:39.709644 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="extract" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.709652 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="extract" Nov 25 19:48:39 crc kubenswrapper[4759]: E1125 19:48:39.709663 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="util" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.709673 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="util" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.709832 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f3e19-ffc6-4a83-83b1-25467cbcbc4c" containerName="extract" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.710562 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.713089 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.713859 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-t76qg" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.721548 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc"] Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.806779 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r96db\" (UniqueName: \"kubernetes.io/projected/979866bf-3d26-40f9-abfb-81aada52864c-kube-api-access-r96db\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.806831 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-webhook-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.806893 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-apiservice-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.908082 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r96db\" (UniqueName: \"kubernetes.io/projected/979866bf-3d26-40f9-abfb-81aada52864c-kube-api-access-r96db\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.908123 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-webhook-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.908164 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-apiservice-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.913613 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-webhook-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.915566 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/979866bf-3d26-40f9-abfb-81aada52864c-apiservice-cert\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:39 crc kubenswrapper[4759]: I1125 19:48:39.930689 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r96db\" (UniqueName: \"kubernetes.io/projected/979866bf-3d26-40f9-abfb-81aada52864c-kube-api-access-r96db\") pod \"infra-operator-controller-manager-59d5f8f59b-7hjcc\" (UID: \"979866bf-3d26-40f9-abfb-81aada52864c\") " pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:40 crc kubenswrapper[4759]: I1125 19:48:40.032564 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:40 crc kubenswrapper[4759]: I1125 19:48:40.276680 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc"] Nov 25 19:48:40 crc kubenswrapper[4759]: I1125 19:48:40.972019 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" event={"ID":"979866bf-3d26-40f9-abfb-81aada52864c","Type":"ContainerStarted","Data":"10d7f491b59ec1147b7b94e6a74c111c25fa54d028958670a8a05b2f0e543af9"} Nov 25 19:48:41 crc kubenswrapper[4759]: I1125 19:48:41.979634 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" event={"ID":"979866bf-3d26-40f9-abfb-81aada52864c","Type":"ContainerStarted","Data":"c777c695335ee27e8dd7da03f4c4ac728228eb14b177e5fa762488cbbafad947"} Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.006860 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" event={"ID":"979866bf-3d26-40f9-abfb-81aada52864c","Type":"ContainerStarted","Data":"3eb9918020dafd8391edc853201e34b7e5e76409f3346c5aba06d8e8c0015861"} Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.007775 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.028175 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" podStartSLOduration=2.42539958 podStartE2EDuration="6.028158456s" podCreationTimestamp="2025-11-25 19:48:39 +0000 UTC" firstStartedPulling="2025-11-25 19:48:40.292650786 +0000 UTC m=+910.878598487" lastFinishedPulling="2025-11-25 19:48:43.895409662 +0000 UTC m=+914.481357363" observedRunningTime="2025-11-25 19:48:45.025505194 +0000 UTC m=+915.611452915" watchObservedRunningTime="2025-11-25 19:48:45.028158456 +0000 UTC m=+915.614106157" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.201263 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.202267 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.204780 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.204947 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-c8vkl" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.205078 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.205202 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.207905 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.210814 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.223868 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.225265 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.234244 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.235394 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.247980 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.253667 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392025 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392076 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzk5j\" (UniqueName: \"kubernetes.io/projected/3c9d592c-a1e4-4e93-968d-f98c10049b84-kube-api-access-gzk5j\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392097 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kolla-config\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392118 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g862\" (UniqueName: \"kubernetes.io/projected/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kube-api-access-9g862\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392138 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-default\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392288 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htk6k\" (UniqueName: \"kubernetes.io/projected/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kube-api-access-htk6k\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392339 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392365 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392403 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392506 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392550 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392581 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392608 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-kolla-config\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392628 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392693 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-default\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392724 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-default\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392751 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kolla-config\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.392863 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494031 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kolla-config\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494071 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g862\" (UniqueName: \"kubernetes.io/projected/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kube-api-access-9g862\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494089 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-default\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494112 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htk6k\" (UniqueName: \"kubernetes.io/projected/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kube-api-access-htk6k\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494133 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494149 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494167 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494183 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494201 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494219 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494234 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-kolla-config\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494250 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494270 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-default\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494286 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-default\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494305 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kolla-config\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494339 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494358 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.494385 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzk5j\" (UniqueName: \"kubernetes.io/projected/3c9d592c-a1e4-4e93-968d-f98c10049b84-kube-api-access-gzk5j\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.495020 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.495044 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.495235 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.495386 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.495961 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kolla-config\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.496370 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.496547 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kolla-config\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.496868 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.496900 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-config-data-default\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.496869 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.497065 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-config-data-default\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.497640 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-config-data-default\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.497902 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6deae4a9-9a28-48fa-800d-9749ac2dcf09-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.498707 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e10fc44-a367-4da8-9414-9aa286b2d9c1-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.506100 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c9d592c-a1e4-4e93-968d-f98c10049b84-kolla-config\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.511684 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.513663 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.516516 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.524142 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htk6k\" (UniqueName: \"kubernetes.io/projected/6deae4a9-9a28-48fa-800d-9749ac2dcf09-kube-api-access-htk6k\") pod \"openstack-galera-0\" (UID: \"6deae4a9-9a28-48fa-800d-9749ac2dcf09\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.524930 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.526714 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzk5j\" (UniqueName: \"kubernetes.io/projected/3c9d592c-a1e4-4e93-968d-f98c10049b84-kube-api-access-gzk5j\") pod \"openstack-galera-2\" (UID: \"3c9d592c-a1e4-4e93-968d-f98c10049b84\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.528205 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g862\" (UniqueName: \"kubernetes.io/projected/5e10fc44-a367-4da8-9414-9aa286b2d9c1-kube-api-access-9g862\") pod \"openstack-galera-1\" (UID: \"5e10fc44-a367-4da8-9414-9aa286b2d9c1\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.547195 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.559652 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:48:45 crc kubenswrapper[4759]: I1125 19:48:45.841860 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 19:48:46 crc kubenswrapper[4759]: I1125 19:48:46.013092 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"5e10fc44-a367-4da8-9414-9aa286b2d9c1","Type":"ContainerStarted","Data":"8db894ec800f06a0096fc3f8e17efbb2c81ae4fa8b0a79b1d91450c93ceb2349"} Nov 25 19:48:46 crc kubenswrapper[4759]: I1125 19:48:46.020907 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-59d5f8f59b-7hjcc" Nov 25 19:48:46 crc kubenswrapper[4759]: I1125 19:48:46.114796 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 19:48:46 crc kubenswrapper[4759]: W1125 19:48:46.121392 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c9d592c_a1e4_4e93_968d_f98c10049b84.slice/crio-59a251ed2387a97207736de6961b69a14c5becd9300bf477078cb44786bc2751 WatchSource:0}: Error finding container 59a251ed2387a97207736de6961b69a14c5becd9300bf477078cb44786bc2751: Status 404 returned error can't find the container with id 59a251ed2387a97207736de6961b69a14c5becd9300bf477078cb44786bc2751 Nov 25 19:48:46 crc kubenswrapper[4759]: I1125 19:48:46.121472 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 19:48:47 crc kubenswrapper[4759]: I1125 19:48:47.019476 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6deae4a9-9a28-48fa-800d-9749ac2dcf09","Type":"ContainerStarted","Data":"459a6e1a62129a18d2b029265df2872308855bd9e8666123178ae0a2039f3011"} Nov 25 19:48:47 crc kubenswrapper[4759]: I1125 19:48:47.020973 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3c9d592c-a1e4-4e93-968d-f98c10049b84","Type":"ContainerStarted","Data":"59a251ed2387a97207736de6961b69a14c5becd9300bf477078cb44786bc2751"} Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.514625 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.515549 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.518820 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-7nn9s" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.519092 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.520400 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.640156 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-kolla-config\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.640222 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-config-data\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.640239 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxjkl\" (UniqueName: \"kubernetes.io/projected/fb645c67-9f04-4bf4-a350-00cea6872805-kube-api-access-gxjkl\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.741580 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-kolla-config\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.741704 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-config-data\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.741731 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxjkl\" (UniqueName: \"kubernetes.io/projected/fb645c67-9f04-4bf4-a350-00cea6872805-kube-api-access-gxjkl\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.742754 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-kolla-config\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.742850 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb645c67-9f04-4bf4-a350-00cea6872805-config-data\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.759311 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxjkl\" (UniqueName: \"kubernetes.io/projected/fb645c67-9f04-4bf4-a350-00cea6872805-kube-api-access-gxjkl\") pod \"memcached-0\" (UID: \"fb645c67-9f04-4bf4-a350-00cea6872805\") " pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:48 crc kubenswrapper[4759]: I1125 19:48:48.832721 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.441794 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pzj5w"] Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.443535 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.446570 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-wn99p" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.471477 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pzj5w"] Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.588091 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w8ws\" (UniqueName: \"kubernetes.io/projected/59198c27-1433-45fd-b774-7e95d852ee68-kube-api-access-4w8ws\") pod \"rabbitmq-cluster-operator-index-pzj5w\" (UID: \"59198c27-1433-45fd-b774-7e95d852ee68\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.689721 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w8ws\" (UniqueName: \"kubernetes.io/projected/59198c27-1433-45fd-b774-7e95d852ee68-kube-api-access-4w8ws\") pod \"rabbitmq-cluster-operator-index-pzj5w\" (UID: \"59198c27-1433-45fd-b774-7e95d852ee68\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.717748 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w8ws\" (UniqueName: \"kubernetes.io/projected/59198c27-1433-45fd-b774-7e95d852ee68-kube-api-access-4w8ws\") pod \"rabbitmq-cluster-operator-index-pzj5w\" (UID: \"59198c27-1433-45fd-b774-7e95d852ee68\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:48:51 crc kubenswrapper[4759]: I1125 19:48:51.764872 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:48:54 crc kubenswrapper[4759]: I1125 19:48:54.358938 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 19:48:54 crc kubenswrapper[4759]: I1125 19:48:54.658360 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pzj5w"] Nov 25 19:48:54 crc kubenswrapper[4759]: W1125 19:48:54.663876 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59198c27_1433_45fd_b774_7e95d852ee68.slice/crio-8f6401123c28b17333e0dd0de99008b3d042dc89fd7248973c302a565b40ee0f WatchSource:0}: Error finding container 8f6401123c28b17333e0dd0de99008b3d042dc89fd7248973c302a565b40ee0f: Status 404 returned error can't find the container with id 8f6401123c28b17333e0dd0de99008b3d042dc89fd7248973c302a565b40ee0f Nov 25 19:48:55 crc kubenswrapper[4759]: I1125 19:48:55.067480 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3c9d592c-a1e4-4e93-968d-f98c10049b84","Type":"ContainerStarted","Data":"2682fb8a2f0d7396b0452ba7aa463e1b5194b4baa4e2ca99ba62627de96797b8"} Nov 25 19:48:55 crc kubenswrapper[4759]: I1125 19:48:55.069060 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6deae4a9-9a28-48fa-800d-9749ac2dcf09","Type":"ContainerStarted","Data":"a0d7f00bfde38ab2a1f78ea24905c713e0b417bdb2803c6242932b3dc6ab508e"} Nov 25 19:48:55 crc kubenswrapper[4759]: I1125 19:48:55.072631 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"fb645c67-9f04-4bf4-a350-00cea6872805","Type":"ContainerStarted","Data":"9c852225bbb24d2497fb409cf998a623239165b02c53eb4589a15fe135b2dd59"} Nov 25 19:48:55 crc kubenswrapper[4759]: I1125 19:48:55.074746 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"5e10fc44-a367-4da8-9414-9aa286b2d9c1","Type":"ContainerStarted","Data":"161b0799f8612bafdd671c71a5d409969a2a65e3500dddc82fbd57acc1b4944c"} Nov 25 19:48:55 crc kubenswrapper[4759]: I1125 19:48:55.077438 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" event={"ID":"59198c27-1433-45fd-b774-7e95d852ee68","Type":"ContainerStarted","Data":"8f6401123c28b17333e0dd0de99008b3d042dc89fd7248973c302a565b40ee0f"} Nov 25 19:48:58 crc kubenswrapper[4759]: I1125 19:48:58.104295 4759 generic.go:334] "Generic (PLEG): container finished" podID="3c9d592c-a1e4-4e93-968d-f98c10049b84" containerID="2682fb8a2f0d7396b0452ba7aa463e1b5194b4baa4e2ca99ba62627de96797b8" exitCode=0 Nov 25 19:48:58 crc kubenswrapper[4759]: I1125 19:48:58.104400 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3c9d592c-a1e4-4e93-968d-f98c10049b84","Type":"ContainerDied","Data":"2682fb8a2f0d7396b0452ba7aa463e1b5194b4baa4e2ca99ba62627de96797b8"} Nov 25 19:48:58 crc kubenswrapper[4759]: I1125 19:48:58.106909 4759 generic.go:334] "Generic (PLEG): container finished" podID="5e10fc44-a367-4da8-9414-9aa286b2d9c1" containerID="161b0799f8612bafdd671c71a5d409969a2a65e3500dddc82fbd57acc1b4944c" exitCode=0 Nov 25 19:48:58 crc kubenswrapper[4759]: I1125 19:48:58.114437 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"5e10fc44-a367-4da8-9414-9aa286b2d9c1","Type":"ContainerDied","Data":"161b0799f8612bafdd671c71a5d409969a2a65e3500dddc82fbd57acc1b4944c"} Nov 25 19:48:59 crc kubenswrapper[4759]: I1125 19:48:59.113841 4759 generic.go:334] "Generic (PLEG): container finished" podID="6deae4a9-9a28-48fa-800d-9749ac2dcf09" containerID="a0d7f00bfde38ab2a1f78ea24905c713e0b417bdb2803c6242932b3dc6ab508e" exitCode=0 Nov 25 19:48:59 crc kubenswrapper[4759]: I1125 19:48:59.113877 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6deae4a9-9a28-48fa-800d-9749ac2dcf09","Type":"ContainerDied","Data":"a0d7f00bfde38ab2a1f78ea24905c713e0b417bdb2803c6242932b3dc6ab508e"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.125350 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" event={"ID":"59198c27-1433-45fd-b774-7e95d852ee68","Type":"ContainerStarted","Data":"3950b37762f55991ef5b0492933ecb0c8b6a06b9a0cd5cbf060ea3468fc95807"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.129821 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3c9d592c-a1e4-4e93-968d-f98c10049b84","Type":"ContainerStarted","Data":"1487c78cdfa573dd7a992a25232c3332c01f1aa369a1c223daf8333f0024fa59"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.131953 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6deae4a9-9a28-48fa-800d-9749ac2dcf09","Type":"ContainerStarted","Data":"c7e69044ed26fa40545b683d3b808b08097f399d4496b51c6523dec7afccaddc"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.134425 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"fb645c67-9f04-4bf4-a350-00cea6872805","Type":"ContainerStarted","Data":"38f015e7bba7fcab7187e73024c71c2b9fb045c12a289fb62a3c4742787d7a7e"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.135079 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.139628 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"5e10fc44-a367-4da8-9414-9aa286b2d9c1","Type":"ContainerStarted","Data":"9a524ca948037971fb3bb87e0813b753fea210bfa2f203ea38433f85b19b1b96"} Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.170643 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=7.771710316 podStartE2EDuration="16.170622187s" podCreationTimestamp="2025-11-25 19:48:44 +0000 UTC" firstStartedPulling="2025-11-25 19:48:45.840630497 +0000 UTC m=+916.426578188" lastFinishedPulling="2025-11-25 19:48:54.239542358 +0000 UTC m=+924.825490059" observedRunningTime="2025-11-25 19:49:00.163918676 +0000 UTC m=+930.749866377" watchObservedRunningTime="2025-11-25 19:49:00.170622187 +0000 UTC m=+930.756569888" Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.187553 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=8.896603173 podStartE2EDuration="12.187537993s" podCreationTimestamp="2025-11-25 19:48:48 +0000 UTC" firstStartedPulling="2025-11-25 19:48:54.363695955 +0000 UTC m=+924.949643656" lastFinishedPulling="2025-11-25 19:48:57.654630775 +0000 UTC m=+928.240578476" observedRunningTime="2025-11-25 19:49:00.184313945 +0000 UTC m=+930.770261646" watchObservedRunningTime="2025-11-25 19:49:00.187537993 +0000 UTC m=+930.773485684" Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.211995 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" podStartSLOduration=4.725015581 podStartE2EDuration="9.211982291s" podCreationTimestamp="2025-11-25 19:48:51 +0000 UTC" firstStartedPulling="2025-11-25 19:48:54.666307213 +0000 UTC m=+925.252254914" lastFinishedPulling="2025-11-25 19:48:59.153273933 +0000 UTC m=+929.739221624" observedRunningTime="2025-11-25 19:49:00.20784331 +0000 UTC m=+930.793791011" watchObservedRunningTime="2025-11-25 19:49:00.211982291 +0000 UTC m=+930.797929992" Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.232327 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.112049331 podStartE2EDuration="16.23231174s" podCreationTimestamp="2025-11-25 19:48:44 +0000 UTC" firstStartedPulling="2025-11-25 19:48:46.124844419 +0000 UTC m=+916.710792140" lastFinishedPulling="2025-11-25 19:48:54.245106848 +0000 UTC m=+924.831054549" observedRunningTime="2025-11-25 19:49:00.225856165 +0000 UTC m=+930.811803866" watchObservedRunningTime="2025-11-25 19:49:00.23231174 +0000 UTC m=+930.818259441" Nov 25 19:49:00 crc kubenswrapper[4759]: I1125 19:49:00.244714 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=8.110789516 podStartE2EDuration="16.244695024s" podCreationTimestamp="2025-11-25 19:48:44 +0000 UTC" firstStartedPulling="2025-11-25 19:48:46.110231454 +0000 UTC m=+916.696179155" lastFinishedPulling="2025-11-25 19:48:54.244136962 +0000 UTC m=+924.830084663" observedRunningTime="2025-11-25 19:49:00.243288166 +0000 UTC m=+930.829235887" watchObservedRunningTime="2025-11-25 19:49:00.244695024 +0000 UTC m=+930.830642725" Nov 25 19:49:01 crc kubenswrapper[4759]: I1125 19:49:01.765946 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:49:01 crc kubenswrapper[4759]: I1125 19:49:01.766240 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:49:01 crc kubenswrapper[4759]: I1125 19:49:01.799062 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.223746 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.225287 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.245732 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.268053 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7z56\" (UniqueName: \"kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.268150 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.268187 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.369419 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7z56\" (UniqueName: \"kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.369504 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.369522 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.369936 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.370007 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.386514 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7z56\" (UniqueName: \"kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56\") pod \"community-operators-fn6w9\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.525481 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.525549 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.541591 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.547760 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.547973 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.560914 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.560942 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:49:05 crc kubenswrapper[4759]: I1125 19:49:05.982021 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:05 crc kubenswrapper[4759]: W1125 19:49:05.986596 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde6b637d_13a9_44c3_97a4_eb3bc11b874e.slice/crio-8305e84d2066cffcd1d63377f66e49e87c91ace377dd019714e52a42de5b3b6e WatchSource:0}: Error finding container 8305e84d2066cffcd1d63377f66e49e87c91ace377dd019714e52a42de5b3b6e: Status 404 returned error can't find the container with id 8305e84d2066cffcd1d63377f66e49e87c91ace377dd019714e52a42de5b3b6e Nov 25 19:49:06 crc kubenswrapper[4759]: I1125 19:49:06.176643 4759 generic.go:334] "Generic (PLEG): container finished" podID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerID="e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8" exitCode=0 Nov 25 19:49:06 crc kubenswrapper[4759]: I1125 19:49:06.176690 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerDied","Data":"e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8"} Nov 25 19:49:06 crc kubenswrapper[4759]: I1125 19:49:06.176717 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerStarted","Data":"8305e84d2066cffcd1d63377f66e49e87c91ace377dd019714e52a42de5b3b6e"} Nov 25 19:49:07 crc kubenswrapper[4759]: I1125 19:49:07.185235 4759 generic.go:334] "Generic (PLEG): container finished" podID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerID="e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed" exitCode=0 Nov 25 19:49:07 crc kubenswrapper[4759]: I1125 19:49:07.185280 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerDied","Data":"e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed"} Nov 25 19:49:07 crc kubenswrapper[4759]: I1125 19:49:07.775815 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:49:07 crc kubenswrapper[4759]: I1125 19:49:07.865251 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 19:49:08 crc kubenswrapper[4759]: I1125 19:49:08.194584 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerStarted","Data":"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6"} Nov 25 19:49:08 crc kubenswrapper[4759]: I1125 19:49:08.211281 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fn6w9" podStartSLOduration=1.750642829 podStartE2EDuration="3.211260461s" podCreationTimestamp="2025-11-25 19:49:05 +0000 UTC" firstStartedPulling="2025-11-25 19:49:06.178009042 +0000 UTC m=+936.763956743" lastFinishedPulling="2025-11-25 19:49:07.638626654 +0000 UTC m=+938.224574375" observedRunningTime="2025-11-25 19:49:08.209185864 +0000 UTC m=+938.795133595" watchObservedRunningTime="2025-11-25 19:49:08.211260461 +0000 UTC m=+938.797208162" Nov 25 19:49:08 crc kubenswrapper[4759]: I1125 19:49:08.834427 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 19:49:11 crc kubenswrapper[4759]: I1125 19:49:11.799019 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-pzj5w" Nov 25 19:49:15 crc kubenswrapper[4759]: I1125 19:49:15.542602 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:15 crc kubenswrapper[4759]: I1125 19:49:15.542906 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:15 crc kubenswrapper[4759]: I1125 19:49:15.581135 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:15 crc kubenswrapper[4759]: I1125 19:49:15.643010 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="3c9d592c-a1e4-4e93-968d-f98c10049b84" containerName="galera" probeResult="failure" output=< Nov 25 19:49:15 crc kubenswrapper[4759]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 25 19:49:15 crc kubenswrapper[4759]: > Nov 25 19:49:16 crc kubenswrapper[4759]: I1125 19:49:16.275250 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:18 crc kubenswrapper[4759]: I1125 19:49:18.815432 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:18 crc kubenswrapper[4759]: I1125 19:49:18.815673 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fn6w9" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="registry-server" containerID="cri-o://8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6" gracePeriod=2 Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.171509 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.264203 4759 generic.go:334] "Generic (PLEG): container finished" podID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerID="8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6" exitCode=0 Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.264265 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerDied","Data":"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6"} Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.264291 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fn6w9" event={"ID":"de6b637d-13a9-44c3-97a4-eb3bc11b874e","Type":"ContainerDied","Data":"8305e84d2066cffcd1d63377f66e49e87c91ace377dd019714e52a42de5b3b6e"} Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.264308 4759 scope.go:117] "RemoveContainer" containerID="8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.264266 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fn6w9" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.282992 4759 scope.go:117] "RemoveContainer" containerID="e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.287402 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.299784 4759 scope.go:117] "RemoveContainer" containerID="e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.302251 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities\") pod \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.302323 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7z56\" (UniqueName: \"kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56\") pod \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.302389 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content\") pod \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\" (UID: \"de6b637d-13a9-44c3-97a4-eb3bc11b874e\") " Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.304437 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities" (OuterVolumeSpecName: "utilities") pod "de6b637d-13a9-44c3-97a4-eb3bc11b874e" (UID: "de6b637d-13a9-44c3-97a4-eb3bc11b874e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.315163 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56" (OuterVolumeSpecName: "kube-api-access-r7z56") pod "de6b637d-13a9-44c3-97a4-eb3bc11b874e" (UID: "de6b637d-13a9-44c3-97a4-eb3bc11b874e"). InnerVolumeSpecName "kube-api-access-r7z56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.322668 4759 scope.go:117] "RemoveContainer" containerID="8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6" Nov 25 19:49:19 crc kubenswrapper[4759]: E1125 19:49:19.323800 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6\": container with ID starting with 8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6 not found: ID does not exist" containerID="8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.323833 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6"} err="failed to get container status \"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6\": rpc error: code = NotFound desc = could not find container \"8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6\": container with ID starting with 8418a020da53945700d5dba118281129d650bc0eb0bcda40d15e573dc50933f6 not found: ID does not exist" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.323860 4759 scope.go:117] "RemoveContainer" containerID="e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed" Nov 25 19:49:19 crc kubenswrapper[4759]: E1125 19:49:19.324134 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed\": container with ID starting with e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed not found: ID does not exist" containerID="e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.324204 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed"} err="failed to get container status \"e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed\": rpc error: code = NotFound desc = could not find container \"e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed\": container with ID starting with e6b98088529d8bb8388b69d99ee254c8ad646503df16f4d0478ac6f29fa435ed not found: ID does not exist" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.324247 4759 scope.go:117] "RemoveContainer" containerID="e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8" Nov 25 19:49:19 crc kubenswrapper[4759]: E1125 19:49:19.324874 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8\": container with ID starting with e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8 not found: ID does not exist" containerID="e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.324893 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8"} err="failed to get container status \"e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8\": rpc error: code = NotFound desc = could not find container \"e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8\": container with ID starting with e3092211e4673c67b10269e0c81adc2d2231d0d5057116a7f521ebadca3ba3d8 not found: ID does not exist" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.361799 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de6b637d-13a9-44c3-97a4-eb3bc11b874e" (UID: "de6b637d-13a9-44c3-97a4-eb3bc11b874e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.379211 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.403900 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.403936 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7z56\" (UniqueName: \"kubernetes.io/projected/de6b637d-13a9-44c3-97a4-eb3bc11b874e-kube-api-access-r7z56\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.403947 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6b637d-13a9-44c3-97a4-eb3bc11b874e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.597965 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:19 crc kubenswrapper[4759]: I1125 19:49:19.602396 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fn6w9"] Nov 25 19:49:20 crc kubenswrapper[4759]: I1125 19:49:20.119307 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" path="/var/lib/kubelet/pods/de6b637d-13a9-44c3-97a4-eb3bc11b874e/volumes" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.265076 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t"] Nov 25 19:49:21 crc kubenswrapper[4759]: E1125 19:49:21.265349 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="extract-utilities" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.265364 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="extract-utilities" Nov 25 19:49:21 crc kubenswrapper[4759]: E1125 19:49:21.265382 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="registry-server" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.265400 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="registry-server" Nov 25 19:49:21 crc kubenswrapper[4759]: E1125 19:49:21.265418 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="extract-content" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.265426 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="extract-content" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.265597 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6b637d-13a9-44c3-97a4-eb3bc11b874e" containerName="registry-server" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.266629 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.270370 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.278722 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t"] Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.431321 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4k6s\" (UniqueName: \"kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.431382 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.431555 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.532373 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.532501 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4k6s\" (UniqueName: \"kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.532535 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.533089 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.533505 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.553497 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4k6s\" (UniqueName: \"kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:21 crc kubenswrapper[4759]: I1125 19:49:21.600954 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.011903 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t"] Nov 25 19:49:22 crc kubenswrapper[4759]: W1125 19:49:22.017829 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca467d28_bd2c_46a1_a27a_e34145ec978f.slice/crio-55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c WatchSource:0}: Error finding container 55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c: Status 404 returned error can't find the container with id 55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.113479 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.193398 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.297234 4759 generic.go:334] "Generic (PLEG): container finished" podID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerID="adf8dfb17e7a932fd13231913c33c620867f11da0d93281b0105b3fa1669e9b9" exitCode=0 Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.297706 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" event={"ID":"ca467d28-bd2c-46a1-a27a-e34145ec978f","Type":"ContainerDied","Data":"adf8dfb17e7a932fd13231913c33c620867f11da0d93281b0105b3fa1669e9b9"} Nov 25 19:49:22 crc kubenswrapper[4759]: I1125 19:49:22.297768 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" event={"ID":"ca467d28-bd2c-46a1-a27a-e34145ec978f","Type":"ContainerStarted","Data":"55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c"} Nov 25 19:49:23 crc kubenswrapper[4759]: I1125 19:49:23.306507 4759 generic.go:334] "Generic (PLEG): container finished" podID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerID="e7f04b7596faba79c2f0c61ae61e55a50c6c093af6ec4f1b7e48328d887bb5d2" exitCode=0 Nov 25 19:49:23 crc kubenswrapper[4759]: I1125 19:49:23.306788 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" event={"ID":"ca467d28-bd2c-46a1-a27a-e34145ec978f","Type":"ContainerDied","Data":"e7f04b7596faba79c2f0c61ae61e55a50c6c093af6ec4f1b7e48328d887bb5d2"} Nov 25 19:49:24 crc kubenswrapper[4759]: I1125 19:49:24.314712 4759 generic.go:334] "Generic (PLEG): container finished" podID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerID="9f39c20b0b70b42fef863d98601ff47294048411ecb732e52d2e368f0f900584" exitCode=0 Nov 25 19:49:24 crc kubenswrapper[4759]: I1125 19:49:24.314759 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" event={"ID":"ca467d28-bd2c-46a1-a27a-e34145ec978f","Type":"ContainerDied","Data":"9f39c20b0b70b42fef863d98601ff47294048411ecb732e52d2e368f0f900584"} Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.560002 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.687696 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4k6s\" (UniqueName: \"kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s\") pod \"ca467d28-bd2c-46a1-a27a-e34145ec978f\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.687822 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util\") pod \"ca467d28-bd2c-46a1-a27a-e34145ec978f\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.689709 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle\") pod \"ca467d28-bd2c-46a1-a27a-e34145ec978f\" (UID: \"ca467d28-bd2c-46a1-a27a-e34145ec978f\") " Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.690376 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle" (OuterVolumeSpecName: "bundle") pod "ca467d28-bd2c-46a1-a27a-e34145ec978f" (UID: "ca467d28-bd2c-46a1-a27a-e34145ec978f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.693707 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s" (OuterVolumeSpecName: "kube-api-access-z4k6s") pod "ca467d28-bd2c-46a1-a27a-e34145ec978f" (UID: "ca467d28-bd2c-46a1-a27a-e34145ec978f"). InnerVolumeSpecName "kube-api-access-z4k6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.703926 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util" (OuterVolumeSpecName: "util") pod "ca467d28-bd2c-46a1-a27a-e34145ec978f" (UID: "ca467d28-bd2c-46a1-a27a-e34145ec978f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.792017 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4k6s\" (UniqueName: \"kubernetes.io/projected/ca467d28-bd2c-46a1-a27a-e34145ec978f-kube-api-access-z4k6s\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.792049 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:25 crc kubenswrapper[4759]: I1125 19:49:25.792058 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca467d28-bd2c-46a1-a27a-e34145ec978f-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:49:26 crc kubenswrapper[4759]: I1125 19:49:26.331841 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" event={"ID":"ca467d28-bd2c-46a1-a27a-e34145ec978f","Type":"ContainerDied","Data":"55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c"} Nov 25 19:49:26 crc kubenswrapper[4759]: I1125 19:49:26.331891 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55cda75233c87498bf88ed486c1858f48623a2ecadfd9adc05723b799da30d6c" Nov 25 19:49:26 crc kubenswrapper[4759]: I1125 19:49:26.331927 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.065980 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr"] Nov 25 19:49:32 crc kubenswrapper[4759]: E1125 19:49:32.067925 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="extract" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.068049 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="extract" Nov 25 19:49:32 crc kubenswrapper[4759]: E1125 19:49:32.068135 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="util" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.068220 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="util" Nov 25 19:49:32 crc kubenswrapper[4759]: E1125 19:49:32.068307 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="pull" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.068379 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="pull" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.068690 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca467d28-bd2c-46a1-a27a-e34145ec978f" containerName="extract" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.069440 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.072123 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-lh8r6" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.076783 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr"] Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.178497 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwt54\" (UniqueName: \"kubernetes.io/projected/3d48201b-dacc-49bd-9d62-c7dbb87af2fa-kube-api-access-hwt54\") pod \"rabbitmq-cluster-operator-779fc9694b-26bxr\" (UID: \"3d48201b-dacc-49bd-9d62-c7dbb87af2fa\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.280259 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwt54\" (UniqueName: \"kubernetes.io/projected/3d48201b-dacc-49bd-9d62-c7dbb87af2fa-kube-api-access-hwt54\") pod \"rabbitmq-cluster-operator-779fc9694b-26bxr\" (UID: \"3d48201b-dacc-49bd-9d62-c7dbb87af2fa\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.299225 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwt54\" (UniqueName: \"kubernetes.io/projected/3d48201b-dacc-49bd-9d62-c7dbb87af2fa-kube-api-access-hwt54\") pod \"rabbitmq-cluster-operator-779fc9694b-26bxr\" (UID: \"3d48201b-dacc-49bd-9d62-c7dbb87af2fa\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.436634 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" Nov 25 19:49:32 crc kubenswrapper[4759]: I1125 19:49:32.869976 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr"] Nov 25 19:49:32 crc kubenswrapper[4759]: W1125 19:49:32.874281 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3d48201b_dacc_49bd_9d62_c7dbb87af2fa.slice/crio-cfab3ccabf94913c0a327663727a490f5840659148c2539ee6e1c470e4848666 WatchSource:0}: Error finding container cfab3ccabf94913c0a327663727a490f5840659148c2539ee6e1c470e4848666: Status 404 returned error can't find the container with id cfab3ccabf94913c0a327663727a490f5840659148c2539ee6e1c470e4848666 Nov 25 19:49:33 crc kubenswrapper[4759]: I1125 19:49:33.371617 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" event={"ID":"3d48201b-dacc-49bd-9d62-c7dbb87af2fa","Type":"ContainerStarted","Data":"cfab3ccabf94913c0a327663727a490f5840659148c2539ee6e1c470e4848666"} Nov 25 19:49:36 crc kubenswrapper[4759]: I1125 19:49:36.402989 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" event={"ID":"3d48201b-dacc-49bd-9d62-c7dbb87af2fa","Type":"ContainerStarted","Data":"f9c1047a7c3739d0ff80bcdeeef8ee8580894ffadcfdb55dfc68e688463a6d09"} Nov 25 19:49:36 crc kubenswrapper[4759]: I1125 19:49:36.423505 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-26bxr" podStartSLOduration=1.426947396 podStartE2EDuration="4.423482281s" podCreationTimestamp="2025-11-25 19:49:32 +0000 UTC" firstStartedPulling="2025-11-25 19:49:32.876388154 +0000 UTC m=+963.462335855" lastFinishedPulling="2025-11-25 19:49:35.872923039 +0000 UTC m=+966.458870740" observedRunningTime="2025-11-25 19:49:36.419371849 +0000 UTC m=+967.005319570" watchObservedRunningTime="2025-11-25 19:49:36.423482281 +0000 UTC m=+967.009429982" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.950776 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.951993 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.953388 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.953700 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.953704 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-p8gfz" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.954399 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.954426 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 19:49:40 crc kubenswrapper[4759]: I1125 19:49:40.964750 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.010764 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.010843 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffltw\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-kube-api-access-ffltw\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011042 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011077 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011117 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011170 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011209 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.011267 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.111981 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112042 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112081 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffltw\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-kube-api-access-ffltw\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112145 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112168 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112203 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112240 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112266 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112789 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.112844 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.113029 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.115762 4759 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.115792 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d493139ec240b408535ee6b37403759d8f9414e453e368c794c497a0bd6fc467/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.119397 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.128156 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.128567 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.136261 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffltw\" (UniqueName: \"kubernetes.io/projected/de35a78c-a4e0-4610-89a3-0802b6bb3dfd-kube-api-access-ffltw\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.149864 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c276ce14-e5f7-44a3-a56e-5f15dfec6b1b\") pod \"rabbitmq-server-0\" (UID: \"de35a78c-a4e0-4610-89a3-0802b6bb3dfd\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.270652 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:49:41 crc kubenswrapper[4759]: I1125 19:49:41.666493 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.442336 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"de35a78c-a4e0-4610-89a3-0802b6bb3dfd","Type":"ContainerStarted","Data":"474d7a0094f2f9486874149f8a5fb36b173d475f13d7b6fbd2837da2ff14d8dc"} Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.833524 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-jhrcq"] Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.834389 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.846463 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-4kxw9" Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.849138 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-jhrcq"] Nov 25 19:49:42 crc kubenswrapper[4759]: I1125 19:49:42.950574 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbwtq\" (UniqueName: \"kubernetes.io/projected/02b44d5e-17d3-41e5-91fb-c5bae66cf31c-kube-api-access-kbwtq\") pod \"keystone-operator-index-jhrcq\" (UID: \"02b44d5e-17d3-41e5-91fb-c5bae66cf31c\") " pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:43 crc kubenswrapper[4759]: I1125 19:49:43.051727 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbwtq\" (UniqueName: \"kubernetes.io/projected/02b44d5e-17d3-41e5-91fb-c5bae66cf31c-kube-api-access-kbwtq\") pod \"keystone-operator-index-jhrcq\" (UID: \"02b44d5e-17d3-41e5-91fb-c5bae66cf31c\") " pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:43 crc kubenswrapper[4759]: I1125 19:49:43.080760 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbwtq\" (UniqueName: \"kubernetes.io/projected/02b44d5e-17d3-41e5-91fb-c5bae66cf31c-kube-api-access-kbwtq\") pod \"keystone-operator-index-jhrcq\" (UID: \"02b44d5e-17d3-41e5-91fb-c5bae66cf31c\") " pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:43 crc kubenswrapper[4759]: I1125 19:49:43.170729 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:43 crc kubenswrapper[4759]: I1125 19:49:43.613813 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-jhrcq"] Nov 25 19:49:44 crc kubenswrapper[4759]: I1125 19:49:44.455082 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jhrcq" event={"ID":"02b44d5e-17d3-41e5-91fb-c5bae66cf31c","Type":"ContainerStarted","Data":"07652ba2aafc10e7ae2c53be5d22c3e4ab32c8a31fa99e4d5b50e2bbfdaf1215"} Nov 25 19:49:48 crc kubenswrapper[4759]: I1125 19:49:48.482882 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jhrcq" event={"ID":"02b44d5e-17d3-41e5-91fb-c5bae66cf31c","Type":"ContainerStarted","Data":"5bd69903eb6992ab221a59edaf4fd3ef6633d81fa5b4117a8b4dda7dcc7a0e1b"} Nov 25 19:49:49 crc kubenswrapper[4759]: I1125 19:49:49.496021 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"de35a78c-a4e0-4610-89a3-0802b6bb3dfd","Type":"ContainerStarted","Data":"3ae3ee92ee485a4a6b9eaf80e9b1f8ec5ec47c1a86e32c3f79a83e0368f57ac3"} Nov 25 19:49:49 crc kubenswrapper[4759]: I1125 19:49:49.531885 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-jhrcq" podStartSLOduration=3.824399985 podStartE2EDuration="7.531844362s" podCreationTimestamp="2025-11-25 19:49:42 +0000 UTC" firstStartedPulling="2025-11-25 19:49:43.634703346 +0000 UTC m=+974.220651047" lastFinishedPulling="2025-11-25 19:49:47.342147683 +0000 UTC m=+977.928095424" observedRunningTime="2025-11-25 19:49:48.505519498 +0000 UTC m=+979.091467239" watchObservedRunningTime="2025-11-25 19:49:49.531844362 +0000 UTC m=+980.117792093" Nov 25 19:49:53 crc kubenswrapper[4759]: I1125 19:49:53.171603 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:53 crc kubenswrapper[4759]: I1125 19:49:53.172662 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:53 crc kubenswrapper[4759]: I1125 19:49:53.197304 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:53 crc kubenswrapper[4759]: I1125 19:49:53.553840 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-jhrcq" Nov 25 19:49:54 crc kubenswrapper[4759]: I1125 19:49:54.896194 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:49:54 crc kubenswrapper[4759]: I1125 19:49:54.896635 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.860165 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv"] Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.861307 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.869002 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.869677 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv"] Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.974109 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.974198 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb7v6\" (UniqueName: \"kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:55 crc kubenswrapper[4759]: I1125 19:49:55.974221 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.075319 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb7v6\" (UniqueName: \"kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.075365 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.075421 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.075926 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.076301 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.093673 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb7v6\" (UniqueName: \"kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.218128 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.446804 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv"] Nov 25 19:49:56 crc kubenswrapper[4759]: I1125 19:49:56.551618 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" event={"ID":"a22e732e-34e5-4d03-b491-72522fb7f9e4","Type":"ContainerStarted","Data":"4630ef5916ba818ddacd882f9151ff770ad642ae4a376f8a4af163d6ffc45bbf"} Nov 25 19:49:57 crc kubenswrapper[4759]: I1125 19:49:57.558970 4759 generic.go:334] "Generic (PLEG): container finished" podID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerID="0e4fc9c00522553dba1b889f3a345cbbccb2b69c066719dcc8b29d093bcb095f" exitCode=0 Nov 25 19:49:57 crc kubenswrapper[4759]: I1125 19:49:57.559012 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" event={"ID":"a22e732e-34e5-4d03-b491-72522fb7f9e4","Type":"ContainerDied","Data":"0e4fc9c00522553dba1b889f3a345cbbccb2b69c066719dcc8b29d093bcb095f"} Nov 25 19:49:58 crc kubenswrapper[4759]: I1125 19:49:58.566576 4759 generic.go:334] "Generic (PLEG): container finished" podID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerID="522af00e5ab21f7bcd2f4f19b1a996b693f4b57275252bfb55f90812a1b0196a" exitCode=0 Nov 25 19:49:58 crc kubenswrapper[4759]: I1125 19:49:58.566717 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" event={"ID":"a22e732e-34e5-4d03-b491-72522fb7f9e4","Type":"ContainerDied","Data":"522af00e5ab21f7bcd2f4f19b1a996b693f4b57275252bfb55f90812a1b0196a"} Nov 25 19:49:59 crc kubenswrapper[4759]: I1125 19:49:59.579950 4759 generic.go:334] "Generic (PLEG): container finished" podID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerID="9c06c0596e93d10d748417bf9177886cd3ab172f84d5d8629133e7fb8a7c0d10" exitCode=0 Nov 25 19:49:59 crc kubenswrapper[4759]: I1125 19:49:59.580023 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" event={"ID":"a22e732e-34e5-4d03-b491-72522fb7f9e4","Type":"ContainerDied","Data":"9c06c0596e93d10d748417bf9177886cd3ab172f84d5d8629133e7fb8a7c0d10"} Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.852886 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.946640 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle\") pod \"a22e732e-34e5-4d03-b491-72522fb7f9e4\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.946810 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb7v6\" (UniqueName: \"kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6\") pod \"a22e732e-34e5-4d03-b491-72522fb7f9e4\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.946865 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util\") pod \"a22e732e-34e5-4d03-b491-72522fb7f9e4\" (UID: \"a22e732e-34e5-4d03-b491-72522fb7f9e4\") " Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.947730 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle" (OuterVolumeSpecName: "bundle") pod "a22e732e-34e5-4d03-b491-72522fb7f9e4" (UID: "a22e732e-34e5-4d03-b491-72522fb7f9e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.955590 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6" (OuterVolumeSpecName: "kube-api-access-xb7v6") pod "a22e732e-34e5-4d03-b491-72522fb7f9e4" (UID: "a22e732e-34e5-4d03-b491-72522fb7f9e4"). InnerVolumeSpecName "kube-api-access-xb7v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:00 crc kubenswrapper[4759]: I1125 19:50:00.966720 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util" (OuterVolumeSpecName: "util") pod "a22e732e-34e5-4d03-b491-72522fb7f9e4" (UID: "a22e732e-34e5-4d03-b491-72522fb7f9e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.048385 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.048425 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb7v6\" (UniqueName: \"kubernetes.io/projected/a22e732e-34e5-4d03-b491-72522fb7f9e4-kube-api-access-xb7v6\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.048438 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a22e732e-34e5-4d03-b491-72522fb7f9e4-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.594836 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" event={"ID":"a22e732e-34e5-4d03-b491-72522fb7f9e4","Type":"ContainerDied","Data":"4630ef5916ba818ddacd882f9151ff770ad642ae4a376f8a4af163d6ffc45bbf"} Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.595062 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4630ef5916ba818ddacd882f9151ff770ad642ae4a376f8a4af163d6ffc45bbf" Nov 25 19:50:01 crc kubenswrapper[4759]: I1125 19:50:01.595120 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.233088 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:04 crc kubenswrapper[4759]: E1125 19:50:04.233967 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="util" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.233987 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="util" Nov 25 19:50:04 crc kubenswrapper[4759]: E1125 19:50:04.234004 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="extract" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.234015 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="extract" Nov 25 19:50:04 crc kubenswrapper[4759]: E1125 19:50:04.234030 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="pull" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.234040 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="pull" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.234375 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22e732e-34e5-4d03-b491-72522fb7f9e4" containerName="extract" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.235766 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.246802 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.292097 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.292179 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.292254 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jkxr\" (UniqueName: \"kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.393897 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.393971 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.394041 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jkxr\" (UniqueName: \"kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.394482 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.394809 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.412918 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jkxr\" (UniqueName: \"kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr\") pod \"redhat-marketplace-q5pvx\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.564137 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:04 crc kubenswrapper[4759]: I1125 19:50:04.829611 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:04 crc kubenswrapper[4759]: W1125 19:50:04.842952 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeae011e7_f7b6_4d29_8b8f_422274f1422d.slice/crio-3830b4c6c98ce753be3c5a5be1c849bc37b2ae7e29f70925f38c1c808d16864c WatchSource:0}: Error finding container 3830b4c6c98ce753be3c5a5be1c849bc37b2ae7e29f70925f38c1c808d16864c: Status 404 returned error can't find the container with id 3830b4c6c98ce753be3c5a5be1c849bc37b2ae7e29f70925f38c1c808d16864c Nov 25 19:50:05 crc kubenswrapper[4759]: I1125 19:50:05.621268 4759 generic.go:334] "Generic (PLEG): container finished" podID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerID="fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65" exitCode=0 Nov 25 19:50:05 crc kubenswrapper[4759]: I1125 19:50:05.621375 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerDied","Data":"fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65"} Nov 25 19:50:05 crc kubenswrapper[4759]: I1125 19:50:05.621601 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerStarted","Data":"3830b4c6c98ce753be3c5a5be1c849bc37b2ae7e29f70925f38c1c808d16864c"} Nov 25 19:50:06 crc kubenswrapper[4759]: I1125 19:50:06.629320 4759 generic.go:334] "Generic (PLEG): container finished" podID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerID="6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9" exitCode=0 Nov 25 19:50:06 crc kubenswrapper[4759]: I1125 19:50:06.629416 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerDied","Data":"6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9"} Nov 25 19:50:07 crc kubenswrapper[4759]: I1125 19:50:07.636997 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerStarted","Data":"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581"} Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.471973 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q5pvx" podStartSLOduration=8.10026903 podStartE2EDuration="9.471949331s" podCreationTimestamp="2025-11-25 19:50:04 +0000 UTC" firstStartedPulling="2025-11-25 19:50:05.622711813 +0000 UTC m=+996.208659514" lastFinishedPulling="2025-11-25 19:50:06.994392114 +0000 UTC m=+997.580339815" observedRunningTime="2025-11-25 19:50:07.656968421 +0000 UTC m=+998.242916122" watchObservedRunningTime="2025-11-25 19:50:13.471949331 +0000 UTC m=+1004.057897032" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.477495 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg"] Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.478898 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.490792 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg"] Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.494210 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.494734 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-b2dqw" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.523677 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-webhook-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.523792 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk59b\" (UniqueName: \"kubernetes.io/projected/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-kube-api-access-sk59b\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.523830 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-apiservice-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.625138 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-apiservice-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.625199 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-webhook-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.625261 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk59b\" (UniqueName: \"kubernetes.io/projected/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-kube-api-access-sk59b\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.631208 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-apiservice-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.638057 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-webhook-cert\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.641025 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk59b\" (UniqueName: \"kubernetes.io/projected/19f1aaf0-6525-4ea2-96fa-8dc16e3e2871-kube-api-access-sk59b\") pod \"keystone-operator-controller-manager-f964ccf8d-kwvtg\" (UID: \"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871\") " pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:13 crc kubenswrapper[4759]: I1125 19:50:13.808136 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.203239 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg"] Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.565178 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.565233 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.624416 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.684422 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" event={"ID":"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871","Type":"ContainerStarted","Data":"70b4f1d8a1aeac2d41a4966d56df02417ccb99c664d7005df71a0cc0a901d7d7"} Nov 25 19:50:14 crc kubenswrapper[4759]: I1125 19:50:14.735489 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:15 crc kubenswrapper[4759]: I1125 19:50:15.815014 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.420548 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.422114 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.435771 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.470507 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.470674 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xltgc\" (UniqueName: \"kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.470708 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.571761 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.571804 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xltgc\" (UniqueName: \"kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.571833 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.572415 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.572554 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.593513 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xltgc\" (UniqueName: \"kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc\") pod \"redhat-operators-stq2k\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.695154 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q5pvx" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="registry-server" containerID="cri-o://c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581" gracePeriod=2 Nov 25 19:50:16 crc kubenswrapper[4759]: I1125 19:50:16.748215 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.600277 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.674834 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.686366 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jkxr\" (UniqueName: \"kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr\") pod \"eae011e7-f7b6-4d29-8b8f-422274f1422d\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.686494 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content\") pod \"eae011e7-f7b6-4d29-8b8f-422274f1422d\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.686527 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities\") pod \"eae011e7-f7b6-4d29-8b8f-422274f1422d\" (UID: \"eae011e7-f7b6-4d29-8b8f-422274f1422d\") " Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.687481 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities" (OuterVolumeSpecName: "utilities") pod "eae011e7-f7b6-4d29-8b8f-422274f1422d" (UID: "eae011e7-f7b6-4d29-8b8f-422274f1422d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.691535 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr" (OuterVolumeSpecName: "kube-api-access-8jkxr") pod "eae011e7-f7b6-4d29-8b8f-422274f1422d" (UID: "eae011e7-f7b6-4d29-8b8f-422274f1422d"). InnerVolumeSpecName "kube-api-access-8jkxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.702839 4759 generic.go:334] "Generic (PLEG): container finished" podID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerID="c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581" exitCode=0 Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.702880 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerDied","Data":"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581"} Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.702922 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5pvx" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.702931 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5pvx" event={"ID":"eae011e7-f7b6-4d29-8b8f-422274f1422d","Type":"ContainerDied","Data":"3830b4c6c98ce753be3c5a5be1c849bc37b2ae7e29f70925f38c1c808d16864c"} Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.702953 4759 scope.go:117] "RemoveContainer" containerID="c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.704206 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" event={"ID":"19f1aaf0-6525-4ea2-96fa-8dc16e3e2871","Type":"ContainerStarted","Data":"86ea9adb305adb0d06412404974c1179bdd13a5bc6210497436d45a639d320b2"} Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.704727 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.708271 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerStarted","Data":"ac264151aabe5b12e8fde9a3d713a0c4a57cf26a7b0eb23c2603f8583e41c13a"} Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.715000 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eae011e7-f7b6-4d29-8b8f-422274f1422d" (UID: "eae011e7-f7b6-4d29-8b8f-422274f1422d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.738389 4759 scope.go:117] "RemoveContainer" containerID="6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.744036 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" podStartSLOduration=1.5798570779999999 podStartE2EDuration="4.74401703s" podCreationTimestamp="2025-11-25 19:50:13 +0000 UTC" firstStartedPulling="2025-11-25 19:50:14.20970401 +0000 UTC m=+1004.795651711" lastFinishedPulling="2025-11-25 19:50:17.373863961 +0000 UTC m=+1007.959811663" observedRunningTime="2025-11-25 19:50:17.732677348 +0000 UTC m=+1008.318625059" watchObservedRunningTime="2025-11-25 19:50:17.74401703 +0000 UTC m=+1008.329964731" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.757877 4759 scope.go:117] "RemoveContainer" containerID="fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.781718 4759 scope.go:117] "RemoveContainer" containerID="c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581" Nov 25 19:50:17 crc kubenswrapper[4759]: E1125 19:50:17.782086 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581\": container with ID starting with c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581 not found: ID does not exist" containerID="c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.782132 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581"} err="failed to get container status \"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581\": rpc error: code = NotFound desc = could not find container \"c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581\": container with ID starting with c4ea589707e5d2cfa16af8f5af058ee455705566a8f66ed67397407a05222581 not found: ID does not exist" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.782160 4759 scope.go:117] "RemoveContainer" containerID="6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9" Nov 25 19:50:17 crc kubenswrapper[4759]: E1125 19:50:17.782464 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9\": container with ID starting with 6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9 not found: ID does not exist" containerID="6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.782496 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9"} err="failed to get container status \"6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9\": rpc error: code = NotFound desc = could not find container \"6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9\": container with ID starting with 6fd52616996f92fd3a4f11738e8579e706ed863f4df8a3c69579c7ce0a7530b9 not found: ID does not exist" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.782523 4759 scope.go:117] "RemoveContainer" containerID="fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65" Nov 25 19:50:17 crc kubenswrapper[4759]: E1125 19:50:17.785188 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65\": container with ID starting with fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65 not found: ID does not exist" containerID="fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.785216 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65"} err="failed to get container status \"fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65\": rpc error: code = NotFound desc = could not find container \"fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65\": container with ID starting with fc97f4894ad6cac26d25c0d5990268d5121db1ac83b70404c07ff3dc70efaa65 not found: ID does not exist" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.787999 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jkxr\" (UniqueName: \"kubernetes.io/projected/eae011e7-f7b6-4d29-8b8f-422274f1422d-kube-api-access-8jkxr\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.788024 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:17 crc kubenswrapper[4759]: I1125 19:50:17.788034 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae011e7-f7b6-4d29-8b8f-422274f1422d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:18 crc kubenswrapper[4759]: I1125 19:50:18.026884 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:18 crc kubenswrapper[4759]: I1125 19:50:18.031217 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5pvx"] Nov 25 19:50:18 crc kubenswrapper[4759]: I1125 19:50:18.116282 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" path="/var/lib/kubelet/pods/eae011e7-f7b6-4d29-8b8f-422274f1422d/volumes" Nov 25 19:50:18 crc kubenswrapper[4759]: I1125 19:50:18.714896 4759 generic.go:334] "Generic (PLEG): container finished" podID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerID="17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368" exitCode=0 Nov 25 19:50:18 crc kubenswrapper[4759]: I1125 19:50:18.714967 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerDied","Data":"17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368"} Nov 25 19:50:19 crc kubenswrapper[4759]: I1125 19:50:19.731944 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerStarted","Data":"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8"} Nov 25 19:50:20 crc kubenswrapper[4759]: I1125 19:50:20.754008 4759 generic.go:334] "Generic (PLEG): container finished" podID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerID="e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8" exitCode=0 Nov 25 19:50:20 crc kubenswrapper[4759]: I1125 19:50:20.754071 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerDied","Data":"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8"} Nov 25 19:50:21 crc kubenswrapper[4759]: I1125 19:50:21.761540 4759 generic.go:334] "Generic (PLEG): container finished" podID="de35a78c-a4e0-4610-89a3-0802b6bb3dfd" containerID="3ae3ee92ee485a4a6b9eaf80e9b1f8ec5ec47c1a86e32c3f79a83e0368f57ac3" exitCode=0 Nov 25 19:50:21 crc kubenswrapper[4759]: I1125 19:50:21.761616 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"de35a78c-a4e0-4610-89a3-0802b6bb3dfd","Type":"ContainerDied","Data":"3ae3ee92ee485a4a6b9eaf80e9b1f8ec5ec47c1a86e32c3f79a83e0368f57ac3"} Nov 25 19:50:21 crc kubenswrapper[4759]: I1125 19:50:21.764832 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerStarted","Data":"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e"} Nov 25 19:50:21 crc kubenswrapper[4759]: I1125 19:50:21.815635 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-stq2k" podStartSLOduration=3.369137742 podStartE2EDuration="5.815605628s" podCreationTimestamp="2025-11-25 19:50:16 +0000 UTC" firstStartedPulling="2025-11-25 19:50:18.716305795 +0000 UTC m=+1009.302253496" lastFinishedPulling="2025-11-25 19:50:21.162773681 +0000 UTC m=+1011.748721382" observedRunningTime="2025-11-25 19:50:21.803129016 +0000 UTC m=+1012.389076747" watchObservedRunningTime="2025-11-25 19:50:21.815605628 +0000 UTC m=+1012.401553369" Nov 25 19:50:22 crc kubenswrapper[4759]: I1125 19:50:22.775375 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"de35a78c-a4e0-4610-89a3-0802b6bb3dfd","Type":"ContainerStarted","Data":"9bd6665a95173c3870a14c5935882168b4223f1371e7cdb70644464991b617bc"} Nov 25 19:50:22 crc kubenswrapper[4759]: I1125 19:50:22.776031 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:50:22 crc kubenswrapper[4759]: I1125 19:50:22.802942 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.42457012 podStartE2EDuration="43.802925324s" podCreationTimestamp="2025-11-25 19:49:39 +0000 UTC" firstStartedPulling="2025-11-25 19:49:41.677679752 +0000 UTC m=+972.263627463" lastFinishedPulling="2025-11-25 19:49:48.056034976 +0000 UTC m=+978.641982667" observedRunningTime="2025-11-25 19:50:22.798803914 +0000 UTC m=+1013.384751615" watchObservedRunningTime="2025-11-25 19:50:22.802925324 +0000 UTC m=+1013.388873025" Nov 25 19:50:23 crc kubenswrapper[4759]: I1125 19:50:23.812933 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f964ccf8d-kwvtg" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.868762 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-hc6nr"] Nov 25 19:50:24 crc kubenswrapper[4759]: E1125 19:50:24.869320 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="registry-server" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.869332 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="registry-server" Nov 25 19:50:24 crc kubenswrapper[4759]: E1125 19:50:24.869353 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="extract-utilities" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.869359 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="extract-utilities" Nov 25 19:50:24 crc kubenswrapper[4759]: E1125 19:50:24.869368 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="extract-content" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.869374 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="extract-content" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.869510 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae011e7-f7b6-4d29-8b8f-422274f1422d" containerName="registry-server" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.869931 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.872689 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr"] Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.873497 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.875748 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.878535 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hc6nr"] Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.882684 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr"] Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.901896 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.901946 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.902478 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.902536 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.902579 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bfqr\" (UniqueName: \"kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:24 crc kubenswrapper[4759]: I1125 19:50:24.902624 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bvm\" (UniqueName: \"kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.003368 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bvm\" (UniqueName: \"kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.003427 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.003502 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.003563 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bfqr\" (UniqueName: \"kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.004271 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.004362 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.021735 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bvm\" (UniqueName: \"kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm\") pod \"keystone-db-create-hc6nr\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.022402 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bfqr\" (UniqueName: \"kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr\") pod \"keystone-ef97-account-create-update-4bbpr\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.185713 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.202694 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.460235 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hc6nr"] Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.590767 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr"] Nov 25 19:50:25 crc kubenswrapper[4759]: W1125 19:50:25.642702 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82702902_862c_4e6e_90b3_6736d46fe003.slice/crio-ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840 WatchSource:0}: Error finding container ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840: Status 404 returned error can't find the container with id ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840 Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.796767 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hc6nr" event={"ID":"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b","Type":"ContainerStarted","Data":"d811a025688663f6c2e09b4f7cdf54533d999d93966f1ff6cb5abdb5f945f1c8"} Nov 25 19:50:25 crc kubenswrapper[4759]: I1125 19:50:25.797438 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" event={"ID":"82702902-862c-4e6e-90b3-6736d46fe003","Type":"ContainerStarted","Data":"ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840"} Nov 25 19:50:26 crc kubenswrapper[4759]: I1125 19:50:26.749364 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:26 crc kubenswrapper[4759]: I1125 19:50:26.749424 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:26 crc kubenswrapper[4759]: I1125 19:50:26.786559 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:26 crc kubenswrapper[4759]: I1125 19:50:26.839777 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:27 crc kubenswrapper[4759]: I1125 19:50:27.813802 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hc6nr" event={"ID":"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b","Type":"ContainerStarted","Data":"833701616d0e27c7b33326731a68364129704ed5b41e662974ecf72c0611b6c0"} Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.631637 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-7647x"] Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.633091 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.637468 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-7647x"] Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.638876 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-fmvcv" Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.776592 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhjr9\" (UniqueName: \"kubernetes.io/projected/85df7f59-3c49-4ece-a44e-c6641ee5da2f-kube-api-access-hhjr9\") pod \"horizon-operator-index-7647x\" (UID: \"85df7f59-3c49-4ece-a44e-c6641ee5da2f\") " pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.821113 4759 generic.go:334] "Generic (PLEG): container finished" podID="82702902-862c-4e6e-90b3-6736d46fe003" containerID="083fba8fd2dc467b2ead6b6e759a3d1895d0095fac384e46c13615a915a6063a" exitCode=0 Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.821669 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" event={"ID":"82702902-862c-4e6e-90b3-6736d46fe003","Type":"ContainerDied","Data":"083fba8fd2dc467b2ead6b6e759a3d1895d0095fac384e46c13615a915a6063a"} Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.824266 4759 generic.go:334] "Generic (PLEG): container finished" podID="beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" containerID="833701616d0e27c7b33326731a68364129704ed5b41e662974ecf72c0611b6c0" exitCode=0 Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.824316 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hc6nr" event={"ID":"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b","Type":"ContainerDied","Data":"833701616d0e27c7b33326731a68364129704ed5b41e662974ecf72c0611b6c0"} Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.877517 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhjr9\" (UniqueName: \"kubernetes.io/projected/85df7f59-3c49-4ece-a44e-c6641ee5da2f-kube-api-access-hhjr9\") pod \"horizon-operator-index-7647x\" (UID: \"85df7f59-3c49-4ece-a44e-c6641ee5da2f\") " pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.897395 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhjr9\" (UniqueName: \"kubernetes.io/projected/85df7f59-3c49-4ece-a44e-c6641ee5da2f-kube-api-access-hhjr9\") pod \"horizon-operator-index-7647x\" (UID: \"85df7f59-3c49-4ece-a44e-c6641ee5da2f\") " pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:28 crc kubenswrapper[4759]: I1125 19:50:28.954913 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.015371 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.015822 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-stq2k" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="registry-server" containerID="cri-o://55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e" gracePeriod=2 Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.373561 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-7647x"] Nov 25 19:50:29 crc kubenswrapper[4759]: W1125 19:50:29.380291 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85df7f59_3c49_4ece_a44e_c6641ee5da2f.slice/crio-97f9ddede61677f0fdb62c7b14658f12a407e55214effb7e1bc0010f8efa0526 WatchSource:0}: Error finding container 97f9ddede61677f0fdb62c7b14658f12a407e55214effb7e1bc0010f8efa0526: Status 404 returned error can't find the container with id 97f9ddede61677f0fdb62c7b14658f12a407e55214effb7e1bc0010f8efa0526 Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.385270 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.488268 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content\") pod \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.488422 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities\") pod \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.488463 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xltgc\" (UniqueName: \"kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc\") pod \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\" (UID: \"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370\") " Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.489466 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities" (OuterVolumeSpecName: "utilities") pod "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" (UID: "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.494864 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc" (OuterVolumeSpecName: "kube-api-access-xltgc") pod "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" (UID: "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370"). InnerVolumeSpecName "kube-api-access-xltgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.581181 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" (UID: "cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.589981 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xltgc\" (UniqueName: \"kubernetes.io/projected/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-kube-api-access-xltgc\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.590015 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.590025 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.834796 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-7647x" event={"ID":"85df7f59-3c49-4ece-a44e-c6641ee5da2f","Type":"ContainerStarted","Data":"97f9ddede61677f0fdb62c7b14658f12a407e55214effb7e1bc0010f8efa0526"} Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.839197 4759 generic.go:334] "Generic (PLEG): container finished" podID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerID="55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e" exitCode=0 Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.839252 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerDied","Data":"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e"} Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.839299 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stq2k" event={"ID":"cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370","Type":"ContainerDied","Data":"ac264151aabe5b12e8fde9a3d713a0c4a57cf26a7b0eb23c2603f8583e41c13a"} Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.839309 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stq2k" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.839322 4759 scope.go:117] "RemoveContainer" containerID="55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.864529 4759 scope.go:117] "RemoveContainer" containerID="e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.889870 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.895723 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-stq2k"] Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.908182 4759 scope.go:117] "RemoveContainer" containerID="17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.983557 4759 scope.go:117] "RemoveContainer" containerID="55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e" Nov 25 19:50:29 crc kubenswrapper[4759]: E1125 19:50:29.983997 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e\": container with ID starting with 55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e not found: ID does not exist" containerID="55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.984024 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e"} err="failed to get container status \"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e\": rpc error: code = NotFound desc = could not find container \"55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e\": container with ID starting with 55bed25c9ab28fda3f30d4b180c187ba0f0fe48804a3d46701e3e1d9e847422e not found: ID does not exist" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.984045 4759 scope.go:117] "RemoveContainer" containerID="e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8" Nov 25 19:50:29 crc kubenswrapper[4759]: E1125 19:50:29.984353 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8\": container with ID starting with e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8 not found: ID does not exist" containerID="e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.984370 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8"} err="failed to get container status \"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8\": rpc error: code = NotFound desc = could not find container \"e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8\": container with ID starting with e81a9756ea84701eb7878b6130283eda0babad5c7f50b097b5c9b267511ee5f8 not found: ID does not exist" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.984384 4759 scope.go:117] "RemoveContainer" containerID="17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368" Nov 25 19:50:29 crc kubenswrapper[4759]: E1125 19:50:29.985051 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368\": container with ID starting with 17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368 not found: ID does not exist" containerID="17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368" Nov 25 19:50:29 crc kubenswrapper[4759]: I1125 19:50:29.985084 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368"} err="failed to get container status \"17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368\": rpc error: code = NotFound desc = could not find container \"17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368\": container with ID starting with 17eec8045be7a4f87c83fcd9893b54a92c0384e2ac34527d532a3e9a83a2d368 not found: ID does not exist" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.138971 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" path="/var/lib/kubelet/pods/cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370/volumes" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.252504 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.290010 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.404417 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58bvm\" (UniqueName: \"kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm\") pod \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.405363 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts\") pod \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\" (UID: \"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b\") " Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.405526 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts\") pod \"82702902-862c-4e6e-90b3-6736d46fe003\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.405707 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bfqr\" (UniqueName: \"kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr\") pod \"82702902-862c-4e6e-90b3-6736d46fe003\" (UID: \"82702902-862c-4e6e-90b3-6736d46fe003\") " Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.406025 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" (UID: "beaa5eb1-4eef-4fb3-a5ef-57fbf473804b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.406064 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82702902-862c-4e6e-90b3-6736d46fe003" (UID: "82702902-862c-4e6e-90b3-6736d46fe003"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.409788 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm" (OuterVolumeSpecName: "kube-api-access-58bvm") pod "beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" (UID: "beaa5eb1-4eef-4fb3-a5ef-57fbf473804b"). InnerVolumeSpecName "kube-api-access-58bvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.410996 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr" (OuterVolumeSpecName: "kube-api-access-5bfqr") pod "82702902-862c-4e6e-90b3-6736d46fe003" (UID: "82702902-862c-4e6e-90b3-6736d46fe003"). InnerVolumeSpecName "kube-api-access-5bfqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.507116 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58bvm\" (UniqueName: \"kubernetes.io/projected/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-kube-api-access-58bvm\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.507159 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.507172 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82702902-862c-4e6e-90b3-6736d46fe003-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.507184 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bfqr\" (UniqueName: \"kubernetes.io/projected/82702902-862c-4e6e-90b3-6736d46fe003-kube-api-access-5bfqr\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.846911 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-hc6nr" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.846926 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-hc6nr" event={"ID":"beaa5eb1-4eef-4fb3-a5ef-57fbf473804b","Type":"ContainerDied","Data":"d811a025688663f6c2e09b4f7cdf54533d999d93966f1ff6cb5abdb5f945f1c8"} Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.847281 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d811a025688663f6c2e09b4f7cdf54533d999d93966f1ff6cb5abdb5f945f1c8" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.848908 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-7647x" event={"ID":"85df7f59-3c49-4ece-a44e-c6641ee5da2f","Type":"ContainerStarted","Data":"9a9dbba4a4580af60e0f4b839306f243f61c7afae1f56fe426947ed304a5bb8c"} Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.851229 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" event={"ID":"82702902-862c-4e6e-90b3-6736d46fe003","Type":"ContainerDied","Data":"ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840"} Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.851263 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecb60ca1840269c7d9618c509c789e2cf9d051d08f584ba0292968e378cb5840" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.851310 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr" Nov 25 19:50:30 crc kubenswrapper[4759]: I1125 19:50:30.864859 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-7647x" podStartSLOduration=2.072961735 podStartE2EDuration="2.864826976s" podCreationTimestamp="2025-11-25 19:50:28 +0000 UTC" firstStartedPulling="2025-11-25 19:50:29.384906131 +0000 UTC m=+1019.970853832" lastFinishedPulling="2025-11-25 19:50:30.176771362 +0000 UTC m=+1020.762719073" observedRunningTime="2025-11-25 19:50:30.864195389 +0000 UTC m=+1021.450143090" watchObservedRunningTime="2025-11-25 19:50:30.864826976 +0000 UTC m=+1021.450774757" Nov 25 19:50:31 crc kubenswrapper[4759]: I1125 19:50:31.274465 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023145 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-2v6hm"] Nov 25 19:50:33 crc kubenswrapper[4759]: E1125 19:50:33.023759 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="extract-utilities" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023797 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="extract-utilities" Nov 25 19:50:33 crc kubenswrapper[4759]: E1125 19:50:33.023812 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" containerName="mariadb-database-create" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023820 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" containerName="mariadb-database-create" Nov 25 19:50:33 crc kubenswrapper[4759]: E1125 19:50:33.023831 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82702902-862c-4e6e-90b3-6736d46fe003" containerName="mariadb-account-create-update" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023839 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="82702902-862c-4e6e-90b3-6736d46fe003" containerName="mariadb-account-create-update" Nov 25 19:50:33 crc kubenswrapper[4759]: E1125 19:50:33.023877 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="registry-server" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023884 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="registry-server" Nov 25 19:50:33 crc kubenswrapper[4759]: E1125 19:50:33.023915 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="extract-content" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.023921 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="extract-content" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.024062 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="82702902-862c-4e6e-90b3-6736d46fe003" containerName="mariadb-account-create-update" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.024079 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd2e3ee4-0f5a-40ab-9e28-8bf63adf8370" containerName="registry-server" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.024091 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" containerName="mariadb-database-create" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.024928 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.027573 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-ckwjp" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.035163 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-2v6hm"] Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.138238 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzhmr\" (UniqueName: \"kubernetes.io/projected/94edb074-48ac-4b2f-ad03-6fe590c96b5c-kube-api-access-zzhmr\") pod \"swift-operator-index-2v6hm\" (UID: \"94edb074-48ac-4b2f-ad03-6fe590c96b5c\") " pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.239476 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzhmr\" (UniqueName: \"kubernetes.io/projected/94edb074-48ac-4b2f-ad03-6fe590c96b5c-kube-api-access-zzhmr\") pod \"swift-operator-index-2v6hm\" (UID: \"94edb074-48ac-4b2f-ad03-6fe590c96b5c\") " pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.257280 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzhmr\" (UniqueName: \"kubernetes.io/projected/94edb074-48ac-4b2f-ad03-6fe590c96b5c-kube-api-access-zzhmr\") pod \"swift-operator-index-2v6hm\" (UID: \"94edb074-48ac-4b2f-ad03-6fe590c96b5c\") " pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.347127 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.581331 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-2v6hm"] Nov 25 19:50:33 crc kubenswrapper[4759]: W1125 19:50:33.583981 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94edb074_48ac_4b2f_ad03_6fe590c96b5c.slice/crio-e43df1a571b558760cc37940501733989b0040024ac2b2a2b4de23264aa9506d WatchSource:0}: Error finding container e43df1a571b558760cc37940501733989b0040024ac2b2a2b4de23264aa9506d: Status 404 returned error can't find the container with id e43df1a571b558760cc37940501733989b0040024ac2b2a2b4de23264aa9506d Nov 25 19:50:33 crc kubenswrapper[4759]: I1125 19:50:33.870277 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-2v6hm" event={"ID":"94edb074-48ac-4b2f-ad03-6fe590c96b5c","Type":"ContainerStarted","Data":"e43df1a571b558760cc37940501733989b0040024ac2b2a2b4de23264aa9506d"} Nov 25 19:50:34 crc kubenswrapper[4759]: I1125 19:50:34.877615 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-2v6hm" event={"ID":"94edb074-48ac-4b2f-ad03-6fe590c96b5c","Type":"ContainerStarted","Data":"b2f5be710be0f5c48cc44da8dff79c960c01a8773df0bf29e1e240ea6d88fab1"} Nov 25 19:50:34 crc kubenswrapper[4759]: I1125 19:50:34.898056 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-2v6hm" podStartSLOduration=1.098901 podStartE2EDuration="1.898025353s" podCreationTimestamp="2025-11-25 19:50:33 +0000 UTC" firstStartedPulling="2025-11-25 19:50:33.585733113 +0000 UTC m=+1024.171680834" lastFinishedPulling="2025-11-25 19:50:34.384857486 +0000 UTC m=+1024.970805187" observedRunningTime="2025-11-25 19:50:34.892628559 +0000 UTC m=+1025.478576270" watchObservedRunningTime="2025-11-25 19:50:34.898025353 +0000 UTC m=+1025.483973054" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.438089 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-d4l86"] Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.439147 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.440675 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.441385 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.441788 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-mh5gd" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.443090 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.454815 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-d4l86"] Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.571086 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4xzz\" (UniqueName: \"kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.571173 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.672820 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.672919 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4xzz\" (UniqueName: \"kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.678700 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.688303 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4xzz\" (UniqueName: \"kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz\") pod \"keystone-db-sync-d4l86\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:35 crc kubenswrapper[4759]: I1125 19:50:35.753643 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:36 crc kubenswrapper[4759]: I1125 19:50:36.139260 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-d4l86"] Nov 25 19:50:36 crc kubenswrapper[4759]: I1125 19:50:36.891359 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-d4l86" event={"ID":"8677c67a-c513-4719-9924-eb2a34bef984","Type":"ContainerStarted","Data":"985935b268a02e7a5fcd9d7bb9a17ef50fb95b92b108e6af024a4365c97c3770"} Nov 25 19:50:38 crc kubenswrapper[4759]: I1125 19:50:38.955947 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:38 crc kubenswrapper[4759]: I1125 19:50:38.956984 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:38 crc kubenswrapper[4759]: I1125 19:50:38.993638 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:39 crc kubenswrapper[4759]: I1125 19:50:39.948805 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-7647x" Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.348772 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.349351 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.374639 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.935470 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-d4l86" event={"ID":"8677c67a-c513-4719-9924-eb2a34bef984","Type":"ContainerStarted","Data":"7d6a59cb9456bfc3f3bc77e31251984b9c7df415b8abc3bc9cee8fb140dcb003"} Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.953667 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-d4l86" podStartSLOduration=2.326967873 podStartE2EDuration="8.953647493s" podCreationTimestamp="2025-11-25 19:50:35 +0000 UTC" firstStartedPulling="2025-11-25 19:50:36.14747568 +0000 UTC m=+1026.733423381" lastFinishedPulling="2025-11-25 19:50:42.77415529 +0000 UTC m=+1033.360103001" observedRunningTime="2025-11-25 19:50:43.951314921 +0000 UTC m=+1034.537262622" watchObservedRunningTime="2025-11-25 19:50:43.953647493 +0000 UTC m=+1034.539595194" Nov 25 19:50:43 crc kubenswrapper[4759]: I1125 19:50:43.968547 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-2v6hm" Nov 25 19:50:45 crc kubenswrapper[4759]: I1125 19:50:45.953708 4759 generic.go:334] "Generic (PLEG): container finished" podID="8677c67a-c513-4719-9924-eb2a34bef984" containerID="7d6a59cb9456bfc3f3bc77e31251984b9c7df415b8abc3bc9cee8fb140dcb003" exitCode=0 Nov 25 19:50:45 crc kubenswrapper[4759]: I1125 19:50:45.953765 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-d4l86" event={"ID":"8677c67a-c513-4719-9924-eb2a34bef984","Type":"ContainerDied","Data":"7d6a59cb9456bfc3f3bc77e31251984b9c7df415b8abc3bc9cee8fb140dcb003"} Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.316926 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.392600 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4xzz\" (UniqueName: \"kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz\") pod \"8677c67a-c513-4719-9924-eb2a34bef984\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.392831 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data\") pod \"8677c67a-c513-4719-9924-eb2a34bef984\" (UID: \"8677c67a-c513-4719-9924-eb2a34bef984\") " Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.402359 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz" (OuterVolumeSpecName: "kube-api-access-k4xzz") pod "8677c67a-c513-4719-9924-eb2a34bef984" (UID: "8677c67a-c513-4719-9924-eb2a34bef984"). InnerVolumeSpecName "kube-api-access-k4xzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.430037 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data" (OuterVolumeSpecName: "config-data") pod "8677c67a-c513-4719-9924-eb2a34bef984" (UID: "8677c67a-c513-4719-9924-eb2a34bef984"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.496299 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4xzz\" (UniqueName: \"kubernetes.io/projected/8677c67a-c513-4719-9924-eb2a34bef984-kube-api-access-k4xzz\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.496760 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677c67a-c513-4719-9924-eb2a34bef984-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.972425 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-d4l86" event={"ID":"8677c67a-c513-4719-9924-eb2a34bef984","Type":"ContainerDied","Data":"985935b268a02e7a5fcd9d7bb9a17ef50fb95b92b108e6af024a4365c97c3770"} Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.972506 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-d4l86" Nov 25 19:50:47 crc kubenswrapper[4759]: I1125 19:50:47.972558 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="985935b268a02e7a5fcd9d7bb9a17ef50fb95b92b108e6af024a4365c97c3770" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.185480 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-g84tx"] Nov 25 19:50:48 crc kubenswrapper[4759]: E1125 19:50:48.186010 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8677c67a-c513-4719-9924-eb2a34bef984" containerName="keystone-db-sync" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.186042 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="8677c67a-c513-4719-9924-eb2a34bef984" containerName="keystone-db-sync" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.186325 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="8677c67a-c513-4719-9924-eb2a34bef984" containerName="keystone-db-sync" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.187216 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.192240 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-g84tx"] Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.193094 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.193435 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.193584 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.193869 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.194387 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-mh5gd" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.313380 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxcz5\" (UniqueName: \"kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.313527 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.313584 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.313657 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.313697 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.416146 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxcz5\" (UniqueName: \"kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.416348 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.416435 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.416671 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.416731 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.421657 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.422048 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.423755 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.434548 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.438191 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxcz5\" (UniqueName: \"kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5\") pod \"keystone-bootstrap-g84tx\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.518698 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.860914 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8"] Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.862396 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.864519 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.871477 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8"] Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.923915 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.924071 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.924281 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxlv6\" (UniqueName: \"kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.957809 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-g84tx"] Nov 25 19:50:48 crc kubenswrapper[4759]: W1125 19:50:48.963636 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb08be9db_6e35_495b_ae8e_89a8deea7695.slice/crio-8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868 WatchSource:0}: Error finding container 8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868: Status 404 returned error can't find the container with id 8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868 Nov 25 19:50:48 crc kubenswrapper[4759]: I1125 19:50:48.981776 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" event={"ID":"b08be9db-6e35-495b-ae8e-89a8deea7695","Type":"ContainerStarted","Data":"8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868"} Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.025847 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxlv6\" (UniqueName: \"kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.025914 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.025954 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.026395 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.028518 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.049291 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxlv6\" (UniqueName: \"kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.181808 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:49 crc kubenswrapper[4759]: W1125 19:50:49.608608 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb160ce8c_b3e2_4934_8f0c_9a1ff3926b0f.slice/crio-15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229 WatchSource:0}: Error finding container 15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229: Status 404 returned error can't find the container with id 15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229 Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.608611 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8"] Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.853250 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n"] Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.854999 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.867527 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n"] Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.949837 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.949890 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wrx\" (UniqueName: \"kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.949937 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.989768 4759 generic.go:334] "Generic (PLEG): container finished" podID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerID="0e180c461a36baf37e8c4e431296c737e7d43fcadc1f00f12f58a50a9f7356c5" exitCode=0 Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.989843 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" event={"ID":"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f","Type":"ContainerDied","Data":"0e180c461a36baf37e8c4e431296c737e7d43fcadc1f00f12f58a50a9f7356c5"} Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.989870 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" event={"ID":"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f","Type":"ContainerStarted","Data":"15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229"} Nov 25 19:50:49 crc kubenswrapper[4759]: I1125 19:50:49.994764 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" event={"ID":"b08be9db-6e35-495b-ae8e-89a8deea7695","Type":"ContainerStarted","Data":"c7c2278dc4cfb0fc558be90b60957d8a2351fdf5aa9bb6bf92ee8a15ff743f1d"} Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.040459 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" podStartSLOduration=2.040430802 podStartE2EDuration="2.040430802s" podCreationTimestamp="2025-11-25 19:50:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:50:50.02114849 +0000 UTC m=+1040.607096211" watchObservedRunningTime="2025-11-25 19:50:50.040430802 +0000 UTC m=+1040.626378503" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.050846 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.050941 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68wrx\" (UniqueName: \"kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.051002 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.051381 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.051705 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.070245 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68wrx\" (UniqueName: \"kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.203786 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:50 crc kubenswrapper[4759]: I1125 19:50:50.672370 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n"] Nov 25 19:50:51 crc kubenswrapper[4759]: I1125 19:50:51.001829 4759 generic.go:334] "Generic (PLEG): container finished" podID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerID="e20b5c3ce19040e9a50f639b8912e9a9b0adeca71821b17fa66be7b587fdff13" exitCode=0 Nov 25 19:50:51 crc kubenswrapper[4759]: I1125 19:50:51.001910 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" event={"ID":"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f","Type":"ContainerDied","Data":"e20b5c3ce19040e9a50f639b8912e9a9b0adeca71821b17fa66be7b587fdff13"} Nov 25 19:50:51 crc kubenswrapper[4759]: I1125 19:50:51.004722 4759 generic.go:334] "Generic (PLEG): container finished" podID="25813c0d-014c-48be-8468-751289d63d85" containerID="56bfb59ac08023960edbbeedc58ce88acb8b9a5fc8511553c3a361da8dc38f68" exitCode=0 Nov 25 19:50:51 crc kubenswrapper[4759]: I1125 19:50:51.004791 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" event={"ID":"25813c0d-014c-48be-8468-751289d63d85","Type":"ContainerDied","Data":"56bfb59ac08023960edbbeedc58ce88acb8b9a5fc8511553c3a361da8dc38f68"} Nov 25 19:50:51 crc kubenswrapper[4759]: I1125 19:50:51.004936 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" event={"ID":"25813c0d-014c-48be-8468-751289d63d85","Type":"ContainerStarted","Data":"97e3a7cc507f6022bc07b91d68c3fc7e84ee2b2b170022aa5c3725f7ae410880"} Nov 25 19:50:52 crc kubenswrapper[4759]: I1125 19:50:52.013996 4759 generic.go:334] "Generic (PLEG): container finished" podID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerID="82bb0d51aae1d91f5a3a918e185da7aff4727ea3e0b219f963045d3142ca5f77" exitCode=0 Nov 25 19:50:52 crc kubenswrapper[4759]: I1125 19:50:52.014073 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" event={"ID":"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f","Type":"ContainerDied","Data":"82bb0d51aae1d91f5a3a918e185da7aff4727ea3e0b219f963045d3142ca5f77"} Nov 25 19:50:52 crc kubenswrapper[4759]: I1125 19:50:52.015513 4759 generic.go:334] "Generic (PLEG): container finished" podID="b08be9db-6e35-495b-ae8e-89a8deea7695" containerID="c7c2278dc4cfb0fc558be90b60957d8a2351fdf5aa9bb6bf92ee8a15ff743f1d" exitCode=0 Nov 25 19:50:52 crc kubenswrapper[4759]: I1125 19:50:52.015538 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" event={"ID":"b08be9db-6e35-495b-ae8e-89a8deea7695","Type":"ContainerDied","Data":"c7c2278dc4cfb0fc558be90b60957d8a2351fdf5aa9bb6bf92ee8a15ff743f1d"} Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.024175 4759 generic.go:334] "Generic (PLEG): container finished" podID="25813c0d-014c-48be-8468-751289d63d85" containerID="2ca81aa58a2b93638aefb380aafb0d48456c833647ebaace66af5669620be54b" exitCode=0 Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.024259 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" event={"ID":"25813c0d-014c-48be-8468-751289d63d85","Type":"ContainerDied","Data":"2ca81aa58a2b93638aefb380aafb0d48456c833647ebaace66af5669620be54b"} Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.391419 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.397652 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498400 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys\") pod \"b08be9db-6e35-495b-ae8e-89a8deea7695\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498499 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util\") pod \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498572 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxcz5\" (UniqueName: \"kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5\") pod \"b08be9db-6e35-495b-ae8e-89a8deea7695\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498589 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxlv6\" (UniqueName: \"kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6\") pod \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498654 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data\") pod \"b08be9db-6e35-495b-ae8e-89a8deea7695\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498691 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts\") pod \"b08be9db-6e35-495b-ae8e-89a8deea7695\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498735 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle\") pod \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\" (UID: \"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.498756 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys\") pod \"b08be9db-6e35-495b-ae8e-89a8deea7695\" (UID: \"b08be9db-6e35-495b-ae8e-89a8deea7695\") " Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.500259 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle" (OuterVolumeSpecName: "bundle") pod "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" (UID: "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.504371 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b08be9db-6e35-495b-ae8e-89a8deea7695" (UID: "b08be9db-6e35-495b-ae8e-89a8deea7695"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.513470 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts" (OuterVolumeSpecName: "scripts") pod "b08be9db-6e35-495b-ae8e-89a8deea7695" (UID: "b08be9db-6e35-495b-ae8e-89a8deea7695"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.514435 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b08be9db-6e35-495b-ae8e-89a8deea7695" (UID: "b08be9db-6e35-495b-ae8e-89a8deea7695"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.517621 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6" (OuterVolumeSpecName: "kube-api-access-qxlv6") pod "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" (UID: "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f"). InnerVolumeSpecName "kube-api-access-qxlv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.517672 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5" (OuterVolumeSpecName: "kube-api-access-lxcz5") pod "b08be9db-6e35-495b-ae8e-89a8deea7695" (UID: "b08be9db-6e35-495b-ae8e-89a8deea7695"). InnerVolumeSpecName "kube-api-access-lxcz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.521347 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util" (OuterVolumeSpecName: "util") pod "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" (UID: "b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.523113 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data" (OuterVolumeSpecName: "config-data") pod "b08be9db-6e35-495b-ae8e-89a8deea7695" (UID: "b08be9db-6e35-495b-ae8e-89a8deea7695"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600094 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600139 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600151 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600163 4759 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600178 4759 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b08be9db-6e35-495b-ae8e-89a8deea7695-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600190 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600201 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxcz5\" (UniqueName: \"kubernetes.io/projected/b08be9db-6e35-495b-ae8e-89a8deea7695-kube-api-access-lxcz5\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:53 crc kubenswrapper[4759]: I1125 19:50:53.600215 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxlv6\" (UniqueName: \"kubernetes.io/projected/b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f-kube-api-access-qxlv6\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.034412 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" event={"ID":"b08be9db-6e35-495b-ae8e-89a8deea7695","Type":"ContainerDied","Data":"8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868"} Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.034516 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8167c491d64b2841d9bb6ae56e79a9727c3cf3f91b4d5141bc6fe2cc7e38b868" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.034612 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-g84tx" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.037126 4759 generic.go:334] "Generic (PLEG): container finished" podID="25813c0d-014c-48be-8468-751289d63d85" containerID="eea73486c505ed6b29d9bba30c84fce8e8611e5c0946ae4b83a5baf4321b87fd" exitCode=0 Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.037216 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" event={"ID":"25813c0d-014c-48be-8468-751289d63d85","Type":"ContainerDied","Data":"eea73486c505ed6b29d9bba30c84fce8e8611e5c0946ae4b83a5baf4321b87fd"} Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.040053 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" event={"ID":"b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f","Type":"ContainerDied","Data":"15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229"} Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.040088 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.040107 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15cb7b05c3b19747eae43ed37d7fc2173a841d32477cdbff26fa1b96f73aa229" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139362 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-5b597bd596-ghgvg"] Nov 25 19:50:54 crc kubenswrapper[4759]: E1125 19:50:54.139618 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="pull" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139631 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="pull" Nov 25 19:50:54 crc kubenswrapper[4759]: E1125 19:50:54.139645 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="extract" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139654 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="extract" Nov 25 19:50:54 crc kubenswrapper[4759]: E1125 19:50:54.139670 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08be9db-6e35-495b-ae8e-89a8deea7695" containerName="keystone-bootstrap" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139677 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08be9db-6e35-495b-ae8e-89a8deea7695" containerName="keystone-bootstrap" Nov 25 19:50:54 crc kubenswrapper[4759]: E1125 19:50:54.139701 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="util" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139710 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="util" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139831 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08be9db-6e35-495b-ae8e-89a8deea7695" containerName="keystone-bootstrap" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.139849 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f" containerName="extract" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.140282 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.142323 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.142602 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-mh5gd" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.142477 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.142872 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.153675 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-5b597bd596-ghgvg"] Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.208826 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrh96\" (UniqueName: \"kubernetes.io/projected/b9a9088f-12be-4169-8bd6-2f641703df0f-kube-api-access-xrh96\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.208944 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-config-data\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.208988 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-fernet-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.209022 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-credential-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.209072 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-scripts\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.310843 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrh96\" (UniqueName: \"kubernetes.io/projected/b9a9088f-12be-4169-8bd6-2f641703df0f-kube-api-access-xrh96\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.310945 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-config-data\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.310977 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-fernet-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.310999 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-credential-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.311043 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-scripts\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.314058 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-credential-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.314510 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-config-data\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.314547 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-scripts\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.314912 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a9088f-12be-4169-8bd6-2f641703df0f-fernet-keys\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.332095 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrh96\" (UniqueName: \"kubernetes.io/projected/b9a9088f-12be-4169-8bd6-2f641703df0f-kube-api-access-xrh96\") pod \"keystone-5b597bd596-ghgvg\" (UID: \"b9a9088f-12be-4169-8bd6-2f641703df0f\") " pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.460287 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.895930 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.895988 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.896034 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.899220 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.899306 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8" gracePeriod=600 Nov 25 19:50:54 crc kubenswrapper[4759]: I1125 19:50:54.906260 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-5b597bd596-ghgvg"] Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.046858 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" event={"ID":"b9a9088f-12be-4169-8bd6-2f641703df0f","Type":"ContainerStarted","Data":"39613ac8579135466bb68b408a1dcabb0502e673344f3a32c7ed25c454a287da"} Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.048735 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8" exitCode=0 Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.048915 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8"} Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.048942 4759 scope.go:117] "RemoveContainer" containerID="ffdca4f197f13075e76d4db8eb985af3c8af079090718249682f7a2544bf1f18" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.265872 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.327085 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util\") pod \"25813c0d-014c-48be-8468-751289d63d85\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.327139 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68wrx\" (UniqueName: \"kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx\") pod \"25813c0d-014c-48be-8468-751289d63d85\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.327207 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle\") pod \"25813c0d-014c-48be-8468-751289d63d85\" (UID: \"25813c0d-014c-48be-8468-751289d63d85\") " Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.328085 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle" (OuterVolumeSpecName: "bundle") pod "25813c0d-014c-48be-8468-751289d63d85" (UID: "25813c0d-014c-48be-8468-751289d63d85"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.338107 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx" (OuterVolumeSpecName: "kube-api-access-68wrx") pod "25813c0d-014c-48be-8468-751289d63d85" (UID: "25813c0d-014c-48be-8468-751289d63d85"). InnerVolumeSpecName "kube-api-access-68wrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.344122 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util" (OuterVolumeSpecName: "util") pod "25813c0d-014c-48be-8468-751289d63d85" (UID: "25813c0d-014c-48be-8468-751289d63d85"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.428575 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.428955 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68wrx\" (UniqueName: \"kubernetes.io/projected/25813c0d-014c-48be-8468-751289d63d85-kube-api-access-68wrx\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:55 crc kubenswrapper[4759]: I1125 19:50:55.428969 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25813c0d-014c-48be-8468-751289d63d85-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.057264 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" event={"ID":"b9a9088f-12be-4169-8bd6-2f641703df0f","Type":"ContainerStarted","Data":"11f0d7ff8bc6de8c19a4f258dffaf101c9b6225b11a5c45ea458ac7952585456"} Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.057362 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.060621 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" event={"ID":"25813c0d-014c-48be-8468-751289d63d85","Type":"ContainerDied","Data":"97e3a7cc507f6022bc07b91d68c3fc7e84ee2b2b170022aa5c3725f7ae410880"} Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.060648 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n" Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.060658 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97e3a7cc507f6022bc07b91d68c3fc7e84ee2b2b170022aa5c3725f7ae410880" Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.063297 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930"} Nov 25 19:50:56 crc kubenswrapper[4759]: I1125 19:50:56.077471 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" podStartSLOduration=2.077428146 podStartE2EDuration="2.077428146s" podCreationTimestamp="2025-11-25 19:50:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:50:56.076107462 +0000 UTC m=+1046.662055193" watchObservedRunningTime="2025-11-25 19:50:56.077428146 +0000 UTC m=+1046.663375857" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.832727 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf"] Nov 25 19:51:13 crc kubenswrapper[4759]: E1125 19:51:13.833563 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="pull" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.833576 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="pull" Nov 25 19:51:13 crc kubenswrapper[4759]: E1125 19:51:13.833586 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="extract" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.834015 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="extract" Nov 25 19:51:13 crc kubenswrapper[4759]: E1125 19:51:13.834037 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="util" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.834043 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="util" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.834199 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="25813c0d-014c-48be-8468-751289d63d85" containerName="extract" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.834681 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.836577 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-l7wkh" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.837168 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.851848 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf"] Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.917772 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-apiservice-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.917821 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-webhook-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:13 crc kubenswrapper[4759]: I1125 19:51:13.918042 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgtt2\" (UniqueName: \"kubernetes.io/projected/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-kube-api-access-zgtt2\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.019407 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgtt2\" (UniqueName: \"kubernetes.io/projected/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-kube-api-access-zgtt2\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.019532 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-apiservice-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.019560 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-webhook-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.029378 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-apiservice-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.029436 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-webhook-cert\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.038532 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgtt2\" (UniqueName: \"kubernetes.io/projected/47df9f04-b8f4-4a25-b578-3d7fe9cb402e-kube-api-access-zgtt2\") pod \"horizon-operator-controller-manager-7444b5967c-rfslf\" (UID: \"47df9f04-b8f4-4a25-b578-3d7fe9cb402e\") " pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.151789 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.699202 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf"] Nov 25 19:51:14 crc kubenswrapper[4759]: I1125 19:51:14.711983 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.196223 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" event={"ID":"47df9f04-b8f4-4a25-b578-3d7fe9cb402e","Type":"ContainerStarted","Data":"670c1d8ec7ec5e12acff49f71f9aacc2e7ddcf8cb23ae8db55dae028882d133d"} Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.902723 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj"] Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.903832 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.905901 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.906155 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-pbxwx" Nov 25 19:51:15 crc kubenswrapper[4759]: I1125 19:51:15.919070 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj"] Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.052152 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-apiservice-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.052337 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-webhook-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.052434 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4pkd\" (UniqueName: \"kubernetes.io/projected/ec15c425-7822-499c-8bea-f9bad0f431a8-kube-api-access-f4pkd\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.154248 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4pkd\" (UniqueName: \"kubernetes.io/projected/ec15c425-7822-499c-8bea-f9bad0f431a8-kube-api-access-f4pkd\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.154852 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-apiservice-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.154965 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-webhook-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.163819 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-apiservice-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.163866 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec15c425-7822-499c-8bea-f9bad0f431a8-webhook-cert\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.174176 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4pkd\" (UniqueName: \"kubernetes.io/projected/ec15c425-7822-499c-8bea-f9bad0f431a8-kube-api-access-f4pkd\") pod \"swift-operator-controller-manager-9975d566b-gc7jj\" (UID: \"ec15c425-7822-499c-8bea-f9bad0f431a8\") " pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:16 crc kubenswrapper[4759]: I1125 19:51:16.232847 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:17 crc kubenswrapper[4759]: I1125 19:51:17.226969 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" event={"ID":"47df9f04-b8f4-4a25-b578-3d7fe9cb402e","Type":"ContainerStarted","Data":"58e29023767efb8de820eff14b8502e7bfe24fab4824f907cba3c58046a99b1f"} Nov 25 19:51:17 crc kubenswrapper[4759]: I1125 19:51:17.227622 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:17 crc kubenswrapper[4759]: I1125 19:51:17.242984 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" podStartSLOduration=2.07718628 podStartE2EDuration="4.242961351s" podCreationTimestamp="2025-11-25 19:51:13 +0000 UTC" firstStartedPulling="2025-11-25 19:51:14.711628324 +0000 UTC m=+1065.297576045" lastFinishedPulling="2025-11-25 19:51:16.877403425 +0000 UTC m=+1067.463351116" observedRunningTime="2025-11-25 19:51:17.241649946 +0000 UTC m=+1067.827597687" watchObservedRunningTime="2025-11-25 19:51:17.242961351 +0000 UTC m=+1067.828909052" Nov 25 19:51:17 crc kubenswrapper[4759]: I1125 19:51:17.301652 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj"] Nov 25 19:51:18 crc kubenswrapper[4759]: I1125 19:51:18.248080 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" event={"ID":"ec15c425-7822-499c-8bea-f9bad0f431a8","Type":"ContainerStarted","Data":"fcdca75a4e047b45147feeb697d22c818aca686c0825be6c61b024501b1618ca"} Nov 25 19:51:20 crc kubenswrapper[4759]: I1125 19:51:20.261858 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" event={"ID":"ec15c425-7822-499c-8bea-f9bad0f431a8","Type":"ContainerStarted","Data":"a2387931ca98d0b14d4c9074cc4a805b07f9f68a5c696f12cea37a40665660e4"} Nov 25 19:51:20 crc kubenswrapper[4759]: I1125 19:51:20.262309 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:20 crc kubenswrapper[4759]: I1125 19:51:20.284394 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" podStartSLOduration=3.391175471 podStartE2EDuration="5.284377733s" podCreationTimestamp="2025-11-25 19:51:15 +0000 UTC" firstStartedPulling="2025-11-25 19:51:17.324870643 +0000 UTC m=+1067.910818344" lastFinishedPulling="2025-11-25 19:51:19.218072905 +0000 UTC m=+1069.804020606" observedRunningTime="2025-11-25 19:51:20.279718869 +0000 UTC m=+1070.865666580" watchObservedRunningTime="2025-11-25 19:51:20.284377733 +0000 UTC m=+1070.870325434" Nov 25 19:51:24 crc kubenswrapper[4759]: I1125 19:51:24.159017 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7444b5967c-rfslf" Nov 25 19:51:25 crc kubenswrapper[4759]: I1125 19:51:25.911182 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-5b597bd596-ghgvg" Nov 25 19:51:26 crc kubenswrapper[4759]: I1125 19:51:26.237108 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9975d566b-gc7jj" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.312883 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.317420 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.319036 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.319100 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.319168 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.328320 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-nk545" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.340219 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.398578 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.398625 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.398651 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-cache\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.398814 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-lock\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.398927 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2fsm\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-kube-api-access-j2fsm\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.500827 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-cache\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.500912 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-lock\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.500959 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2fsm\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-kube-api-access-j2fsm\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.501012 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.501036 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: E1125 19:51:32.501167 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:32 crc kubenswrapper[4759]: E1125 19:51:32.501182 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 19:51:32 crc kubenswrapper[4759]: E1125 19:51:32.501227 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift podName:07cbf89a-8a0b-48ab-9e23-52226f362fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:51:33.001208595 +0000 UTC m=+1083.587156286 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift") pod "swift-storage-0" (UID: "07cbf89a-8a0b-48ab-9e23-52226f362fbc") : configmap "swift-ring-files" not found Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.501783 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.501841 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-lock\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.504899 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/07cbf89a-8a0b-48ab-9e23-52226f362fbc-cache\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.521007 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:32 crc kubenswrapper[4759]: I1125 19:51:32.523573 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2fsm\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-kube-api-access-j2fsm\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.040783 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.041048 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.041068 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.041137 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift podName:07cbf89a-8a0b-48ab-9e23-52226f362fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:51:34.041114815 +0000 UTC m=+1084.627062516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift") pod "swift-storage-0" (UID: "07cbf89a-8a0b-48ab-9e23-52226f362fbc") : configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.576740 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc"] Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.578061 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.579795 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.587016 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc"] Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.647979 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.648266 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln4cr\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-kube-api-access-ln4cr\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.648381 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-config-data\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.648517 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-log-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.648661 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-run-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750255 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln4cr\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-kube-api-access-ln4cr\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750306 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-config-data\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750350 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-log-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750493 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-run-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750571 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.750771 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.750825 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc: configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750878 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-log-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: E1125 19:51:33.750919 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift podName:f4b24bd8-f201-4cbb-8a6c-1821e1c1002c nodeName:}" failed. No retries permitted until 2025-11-25 19:51:34.250858418 +0000 UTC m=+1084.836806139 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift") pod "swift-proxy-6bd58cfcf7-zlfnc" (UID: "f4b24bd8-f201-4cbb-8a6c-1821e1c1002c") : configmap "swift-ring-files" not found Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.750963 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-run-httpd\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.768165 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-config-data\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.785694 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln4cr\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-kube-api-access-ln4cr\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.828897 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.829953 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.831862 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-jdggt" Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.836007 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:33 crc kubenswrapper[4759]: I1125 19:51:33.954169 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5dnc\" (UniqueName: \"kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc\") pod \"glance-operator-index-b6ths\" (UID: \"fa09f5a0-5ab3-449d-91a7-5ab538699479\") " pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.055815 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5dnc\" (UniqueName: \"kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc\") pod \"glance-operator-index-b6ths\" (UID: \"fa09f5a0-5ab3-449d-91a7-5ab538699479\") " pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.056163 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.056324 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.056337 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.056380 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift podName:07cbf89a-8a0b-48ab-9e23-52226f362fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:51:36.056366055 +0000 UTC m=+1086.642313756 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift") pod "swift-storage-0" (UID: "07cbf89a-8a0b-48ab-9e23-52226f362fbc") : configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.074356 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5dnc\" (UniqueName: \"kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc\") pod \"glance-operator-index-b6ths\" (UID: \"fa09f5a0-5ab3-449d-91a7-5ab538699479\") " pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.145863 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.259212 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.259351 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.259363 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc: configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: E1125 19:51:34.259403 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift podName:f4b24bd8-f201-4cbb-8a6c-1821e1c1002c nodeName:}" failed. No retries permitted until 2025-11-25 19:51:35.259389351 +0000 UTC m=+1085.845337052 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift") pod "swift-proxy-6bd58cfcf7-zlfnc" (UID: "f4b24bd8-f201-4cbb-8a6c-1821e1c1002c") : configmap "swift-ring-files" not found Nov 25 19:51:34 crc kubenswrapper[4759]: W1125 19:51:34.684431 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa09f5a0_5ab3_449d_91a7_5ab538699479.slice/crio-b01a3116dbe0b4ca9efd588fd61ae5217737d7e3edaaf5992aabac9a9d5a826c WatchSource:0}: Error finding container b01a3116dbe0b4ca9efd588fd61ae5217737d7e3edaaf5992aabac9a9d5a826c: Status 404 returned error can't find the container with id b01a3116dbe0b4ca9efd588fd61ae5217737d7e3edaaf5992aabac9a9d5a826c Nov 25 19:51:34 crc kubenswrapper[4759]: I1125 19:51:34.686266 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:35 crc kubenswrapper[4759]: I1125 19:51:35.276434 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:35 crc kubenswrapper[4759]: E1125 19:51:35.276678 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:35 crc kubenswrapper[4759]: E1125 19:51:35.276710 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc: configmap "swift-ring-files" not found Nov 25 19:51:35 crc kubenswrapper[4759]: E1125 19:51:35.276770 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift podName:f4b24bd8-f201-4cbb-8a6c-1821e1c1002c nodeName:}" failed. No retries permitted until 2025-11-25 19:51:37.276753327 +0000 UTC m=+1087.862701018 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift") pod "swift-proxy-6bd58cfcf7-zlfnc" (UID: "f4b24bd8-f201-4cbb-8a6c-1821e1c1002c") : configmap "swift-ring-files" not found Nov 25 19:51:35 crc kubenswrapper[4759]: I1125 19:51:35.376112 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-b6ths" event={"ID":"fa09f5a0-5ab3-449d-91a7-5ab538699479","Type":"ContainerStarted","Data":"b01a3116dbe0b4ca9efd588fd61ae5217737d7e3edaaf5992aabac9a9d5a826c"} Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.087415 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:36 crc kubenswrapper[4759]: E1125 19:51:36.087603 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:36 crc kubenswrapper[4759]: E1125 19:51:36.087778 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 19:51:36 crc kubenswrapper[4759]: E1125 19:51:36.087842 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift podName:07cbf89a-8a0b-48ab-9e23-52226f362fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:51:40.087822288 +0000 UTC m=+1090.673769989 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift") pod "swift-storage-0" (UID: "07cbf89a-8a0b-48ab-9e23-52226f362fbc") : configmap "swift-ring-files" not found Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.299148 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vgfs7"] Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.299944 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.302807 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.305508 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.312514 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vgfs7"] Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392609 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392668 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvhn\" (UniqueName: \"kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392772 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392809 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392871 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.392931 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495804 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495849 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495869 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495920 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495940 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvhn\" (UniqueName: \"kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.495994 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.496528 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.497054 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.497148 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.501464 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.501825 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.511060 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvhn\" (UniqueName: \"kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn\") pod \"swift-ring-rebalance-vgfs7\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:36 crc kubenswrapper[4759]: I1125 19:51:36.659710 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:37 crc kubenswrapper[4759]: E1125 19:51:37.321241 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:37 crc kubenswrapper[4759]: E1125 19:51:37.321577 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc: configmap "swift-ring-files" not found Nov 25 19:51:37 crc kubenswrapper[4759]: E1125 19:51:37.321640 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift podName:f4b24bd8-f201-4cbb-8a6c-1821e1c1002c nodeName:}" failed. No retries permitted until 2025-11-25 19:51:41.321619698 +0000 UTC m=+1091.907567399 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift") pod "swift-proxy-6bd58cfcf7-zlfnc" (UID: "f4b24bd8-f201-4cbb-8a6c-1821e1c1002c") : configmap "swift-ring-files" not found Nov 25 19:51:37 crc kubenswrapper[4759]: I1125 19:51:37.321109 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:37 crc kubenswrapper[4759]: I1125 19:51:37.393064 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vgfs7"] Nov 25 19:51:37 crc kubenswrapper[4759]: W1125 19:51:37.701175 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd763ac10_2547_4a41_a788_c784d9a0bd84.slice/crio-9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37 WatchSource:0}: Error finding container 9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37: Status 404 returned error can't find the container with id 9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37 Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.217755 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.402156 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-b6ths" event={"ID":"fa09f5a0-5ab3-449d-91a7-5ab538699479","Type":"ContainerStarted","Data":"51cca95ac9e9705cea5d01a4cf8e50afaaff622f211da6ef3ccbda59c0fdf037"} Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.406501 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" event={"ID":"d763ac10-2547-4a41-a788-c784d9a0bd84","Type":"ContainerStarted","Data":"9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37"} Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.417784 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-b6ths" podStartSLOduration=2.372497557 podStartE2EDuration="5.417764831s" podCreationTimestamp="2025-11-25 19:51:33 +0000 UTC" firstStartedPulling="2025-11-25 19:51:34.686711323 +0000 UTC m=+1085.272659024" lastFinishedPulling="2025-11-25 19:51:37.731978597 +0000 UTC m=+1088.317926298" observedRunningTime="2025-11-25 19:51:38.416240491 +0000 UTC m=+1089.002188192" watchObservedRunningTime="2025-11-25 19:51:38.417764831 +0000 UTC m=+1089.003712532" Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.824535 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-cg7g2"] Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.825634 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.847164 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-cg7g2"] Nov 25 19:51:38 crc kubenswrapper[4759]: I1125 19:51:38.951977 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phrxc\" (UniqueName: \"kubernetes.io/projected/c0e21472-d327-4f63-a8a5-be294881a356-kube-api-access-phrxc\") pod \"glance-operator-index-cg7g2\" (UID: \"c0e21472-d327-4f63-a8a5-be294881a356\") " pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:39 crc kubenswrapper[4759]: I1125 19:51:39.053379 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phrxc\" (UniqueName: \"kubernetes.io/projected/c0e21472-d327-4f63-a8a5-be294881a356-kube-api-access-phrxc\") pod \"glance-operator-index-cg7g2\" (UID: \"c0e21472-d327-4f63-a8a5-be294881a356\") " pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:39 crc kubenswrapper[4759]: I1125 19:51:39.086094 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phrxc\" (UniqueName: \"kubernetes.io/projected/c0e21472-d327-4f63-a8a5-be294881a356-kube-api-access-phrxc\") pod \"glance-operator-index-cg7g2\" (UID: \"c0e21472-d327-4f63-a8a5-be294881a356\") " pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:39 crc kubenswrapper[4759]: I1125 19:51:39.146119 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:39 crc kubenswrapper[4759]: I1125 19:51:39.424977 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-b6ths" podUID="fa09f5a0-5ab3-449d-91a7-5ab538699479" containerName="registry-server" containerID="cri-o://51cca95ac9e9705cea5d01a4cf8e50afaaff622f211da6ef3ccbda59c0fdf037" gracePeriod=2 Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.170579 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:40 crc kubenswrapper[4759]: E1125 19:51:40.170768 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:40 crc kubenswrapper[4759]: E1125 19:51:40.171088 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 19:51:40 crc kubenswrapper[4759]: E1125 19:51:40.171143 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift podName:07cbf89a-8a0b-48ab-9e23-52226f362fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:51:48.171124199 +0000 UTC m=+1098.757071900 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift") pod "swift-storage-0" (UID: "07cbf89a-8a0b-48ab-9e23-52226f362fbc") : configmap "swift-ring-files" not found Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.436528 4759 generic.go:334] "Generic (PLEG): container finished" podID="fa09f5a0-5ab3-449d-91a7-5ab538699479" containerID="51cca95ac9e9705cea5d01a4cf8e50afaaff622f211da6ef3ccbda59c0fdf037" exitCode=0 Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.436570 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-b6ths" event={"ID":"fa09f5a0-5ab3-449d-91a7-5ab538699479","Type":"ContainerDied","Data":"51cca95ac9e9705cea5d01a4cf8e50afaaff622f211da6ef3ccbda59c0fdf037"} Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.694173 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.780631 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5dnc\" (UniqueName: \"kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc\") pod \"fa09f5a0-5ab3-449d-91a7-5ab538699479\" (UID: \"fa09f5a0-5ab3-449d-91a7-5ab538699479\") " Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.784980 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc" (OuterVolumeSpecName: "kube-api-access-d5dnc") pod "fa09f5a0-5ab3-449d-91a7-5ab538699479" (UID: "fa09f5a0-5ab3-449d-91a7-5ab538699479"). InnerVolumeSpecName "kube-api-access-d5dnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.882810 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5dnc\" (UniqueName: \"kubernetes.io/projected/fa09f5a0-5ab3-449d-91a7-5ab538699479-kube-api-access-d5dnc\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:40 crc kubenswrapper[4759]: W1125 19:51:40.958426 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e21472_d327_4f63_a8a5_be294881a356.slice/crio-0c17841fafae0dbdd4585d51b1fac00f5d293ddf39c9b73894c0bd93b13a3ea8 WatchSource:0}: Error finding container 0c17841fafae0dbdd4585d51b1fac00f5d293ddf39c9b73894c0bd93b13a3ea8: Status 404 returned error can't find the container with id 0c17841fafae0dbdd4585d51b1fac00f5d293ddf39c9b73894c0bd93b13a3ea8 Nov 25 19:51:40 crc kubenswrapper[4759]: I1125 19:51:40.963812 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-cg7g2"] Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.390532 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:41 crc kubenswrapper[4759]: E1125 19:51:41.391020 4759 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 19:51:41 crc kubenswrapper[4759]: E1125 19:51:41.391161 4759 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc: configmap "swift-ring-files" not found Nov 25 19:51:41 crc kubenswrapper[4759]: E1125 19:51:41.391224 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift podName:f4b24bd8-f201-4cbb-8a6c-1821e1c1002c nodeName:}" failed. No retries permitted until 2025-11-25 19:51:49.391201894 +0000 UTC m=+1099.977149615 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift") pod "swift-proxy-6bd58cfcf7-zlfnc" (UID: "f4b24bd8-f201-4cbb-8a6c-1821e1c1002c") : configmap "swift-ring-files" not found Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.444132 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-cg7g2" event={"ID":"c0e21472-d327-4f63-a8a5-be294881a356","Type":"ContainerStarted","Data":"a3936fead8198db5eb7f2d1d341bf54ce9cd275ac8282649e0c16fb1cea2fe01"} Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.444181 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-cg7g2" event={"ID":"c0e21472-d327-4f63-a8a5-be294881a356","Type":"ContainerStarted","Data":"0c17841fafae0dbdd4585d51b1fac00f5d293ddf39c9b73894c0bd93b13a3ea8"} Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.446699 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-b6ths" event={"ID":"fa09f5a0-5ab3-449d-91a7-5ab538699479","Type":"ContainerDied","Data":"b01a3116dbe0b4ca9efd588fd61ae5217737d7e3edaaf5992aabac9a9d5a826c"} Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.446746 4759 scope.go:117] "RemoveContainer" containerID="51cca95ac9e9705cea5d01a4cf8e50afaaff622f211da6ef3ccbda59c0fdf037" Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.446874 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-b6ths" Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.453643 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" event={"ID":"d763ac10-2547-4a41-a788-c784d9a0bd84","Type":"ContainerStarted","Data":"9b12cd12ca92132cea116da611eb473335e2b7b7603dc34f18b3e002e085bc39"} Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.494084 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-cg7g2" podStartSLOduration=3.437818676 podStartE2EDuration="3.494065033s" podCreationTimestamp="2025-11-25 19:51:38 +0000 UTC" firstStartedPulling="2025-11-25 19:51:40.963280907 +0000 UTC m=+1091.549228648" lastFinishedPulling="2025-11-25 19:51:41.019527264 +0000 UTC m=+1091.605475005" observedRunningTime="2025-11-25 19:51:41.47702592 +0000 UTC m=+1092.062973621" watchObservedRunningTime="2025-11-25 19:51:41.494065033 +0000 UTC m=+1092.080012734" Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.495330 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" podStartSLOduration=2.60360589 podStartE2EDuration="5.495323566s" podCreationTimestamp="2025-11-25 19:51:36 +0000 UTC" firstStartedPulling="2025-11-25 19:51:37.70319193 +0000 UTC m=+1088.289139631" lastFinishedPulling="2025-11-25 19:51:40.594909606 +0000 UTC m=+1091.180857307" observedRunningTime="2025-11-25 19:51:41.492702397 +0000 UTC m=+1092.078650098" watchObservedRunningTime="2025-11-25 19:51:41.495323566 +0000 UTC m=+1092.081271267" Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.507927 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:41 crc kubenswrapper[4759]: I1125 19:51:41.514422 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-b6ths"] Nov 25 19:51:42 crc kubenswrapper[4759]: I1125 19:51:42.121628 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa09f5a0-5ab3-449d-91a7-5ab538699479" path="/var/lib/kubelet/pods/fa09f5a0-5ab3-449d-91a7-5ab538699479/volumes" Nov 25 19:51:47 crc kubenswrapper[4759]: I1125 19:51:47.497711 4759 generic.go:334] "Generic (PLEG): container finished" podID="d763ac10-2547-4a41-a788-c784d9a0bd84" containerID="9b12cd12ca92132cea116da611eb473335e2b7b7603dc34f18b3e002e085bc39" exitCode=0 Nov 25 19:51:47 crc kubenswrapper[4759]: I1125 19:51:47.497810 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" event={"ID":"d763ac10-2547-4a41-a788-c784d9a0bd84","Type":"ContainerDied","Data":"9b12cd12ca92132cea116da611eb473335e2b7b7603dc34f18b3e002e085bc39"} Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.223833 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.232562 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/07cbf89a-8a0b-48ab-9e23-52226f362fbc-etc-swift\") pod \"swift-storage-0\" (UID: \"07cbf89a-8a0b-48ab-9e23-52226f362fbc\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.532366 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.827855 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934378 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934432 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934516 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934591 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934626 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.934670 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jvhn\" (UniqueName: \"kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn\") pod \"d763ac10-2547-4a41-a788-c784d9a0bd84\" (UID: \"d763ac10-2547-4a41-a788-c784d9a0bd84\") " Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.935400 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.935496 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.939833 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn" (OuterVolumeSpecName: "kube-api-access-2jvhn") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "kube-api-access-2jvhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.950120 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts" (OuterVolumeSpecName: "scripts") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.954608 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:51:48 crc kubenswrapper[4759]: I1125 19:51:48.975843 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d763ac10-2547-4a41-a788-c784d9a0bd84" (UID: "d763ac10-2547-4a41-a788-c784d9a0bd84"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.005908 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 19:51:49 crc kubenswrapper[4759]: W1125 19:51:49.013942 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07cbf89a_8a0b_48ab_9e23_52226f362fbc.slice/crio-1710018b424a84945e77e0e074822b9e3c33274b16995c275f5848736b9aa77a WatchSource:0}: Error finding container 1710018b424a84945e77e0e074822b9e3c33274b16995c275f5848736b9aa77a: Status 404 returned error can't find the container with id 1710018b424a84945e77e0e074822b9e3c33274b16995c275f5848736b9aa77a Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036818 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036853 4759 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d763ac10-2547-4a41-a788-c784d9a0bd84-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036869 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jvhn\" (UniqueName: \"kubernetes.io/projected/d763ac10-2547-4a41-a788-c784d9a0bd84-kube-api-access-2jvhn\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036882 4759 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036896 4759 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d763ac10-2547-4a41-a788-c784d9a0bd84-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.036906 4759 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d763ac10-2547-4a41-a788-c784d9a0bd84-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.146799 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.146855 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.185386 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.443822 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.456126 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f4b24bd8-f201-4cbb-8a6c-1821e1c1002c-etc-swift\") pod \"swift-proxy-6bd58cfcf7-zlfnc\" (UID: \"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.493051 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.527916 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.527986 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-vgfs7" event={"ID":"d763ac10-2547-4a41-a788-c784d9a0bd84","Type":"ContainerDied","Data":"9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37"} Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.528184 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b9b5de2caf4899dde60a4a59471d3a15f1c1a8863e24fff6cf25a4bfffe9e37" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.530634 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"1710018b424a84945e77e0e074822b9e3c33274b16995c275f5848736b9aa77a"} Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.579951 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-cg7g2" Nov 25 19:51:49 crc kubenswrapper[4759]: I1125 19:51:49.943371 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc"] Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.596060 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" event={"ID":"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c","Type":"ContainerStarted","Data":"5265bd9677d7af2bd5ef8d9080a94cc384ef8664c91a4e7497dd0fbfbcf2de8c"} Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.596458 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" event={"ID":"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c","Type":"ContainerStarted","Data":"5bc38840621f602bcf54c13f8189acec17b3f6ca393535a62b1fc414aca5a81a"} Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.596475 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" event={"ID":"f4b24bd8-f201-4cbb-8a6c-1821e1c1002c","Type":"ContainerStarted","Data":"b447184bc832016334e0228ab22c4987692ed9748cf7e0d8b85e7322f7dd8d13"} Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.596492 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.596504 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.621888 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" podStartSLOduration=17.621432394 podStartE2EDuration="17.621432394s" podCreationTimestamp="2025-11-25 19:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:51:50.618818604 +0000 UTC m=+1101.204766315" watchObservedRunningTime="2025-11-25 19:51:50.621432394 +0000 UTC m=+1101.207380105" Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.624013 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"aa7325581134fb3825bd70d8e53baaa15166d732853c66a6a3d77c34800faeae"} Nov 25 19:51:50 crc kubenswrapper[4759]: I1125 19:51:50.624052 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"34efb67244f8ef103bc1881704a03d23d94a1666280252abdd2784cf67aea789"} Nov 25 19:51:51 crc kubenswrapper[4759]: I1125 19:51:51.634612 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"fe6a20c9e1e8cdf7e44790e44939ebb2a8e854cf2efc4578d8b1971bbf0989d4"} Nov 25 19:51:51 crc kubenswrapper[4759]: I1125 19:51:51.634662 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"9d0f7a9c7be4084a2b1464a0520546cd697be50cf16897cb99923b5b8365ad4a"} Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.645480 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"10d811b10799574f11e1f270ea59497c53438f44a8b0f9e23aa96f3d1b00da90"} Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.646022 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"a288bb01b5c02cd8484a598cb04f829e35eb8454e92f1356449afd81e7c7e04e"} Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.646032 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"368808bf5bd2f894e8daf37dea829a96b322797bbab3945d83554e8bdbd2d945"} Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.646040 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"a4a952adaf9400276e6bafdf6cdf19eac85770d29450392d5384fb17b59da18e"} Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.658973 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl"] Nov 25 19:51:52 crc kubenswrapper[4759]: E1125 19:51:52.659225 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d763ac10-2547-4a41-a788-c784d9a0bd84" containerName="swift-ring-rebalance" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.659242 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="d763ac10-2547-4a41-a788-c784d9a0bd84" containerName="swift-ring-rebalance" Nov 25 19:51:52 crc kubenswrapper[4759]: E1125 19:51:52.659254 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa09f5a0-5ab3-449d-91a7-5ab538699479" containerName="registry-server" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.659262 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa09f5a0-5ab3-449d-91a7-5ab538699479" containerName="registry-server" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.659397 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="d763ac10-2547-4a41-a788-c784d9a0bd84" containerName="swift-ring-rebalance" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.659412 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa09f5a0-5ab3-449d-91a7-5ab538699479" containerName="registry-server" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.660307 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.663118 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6wwg6" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.673297 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl"] Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.729229 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.729346 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.729405 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xvd\" (UniqueName: \"kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.831111 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.831160 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.831190 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xvd\" (UniqueName: \"kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.831794 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.831794 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.858512 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xvd\" (UniqueName: \"kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd\") pod \"3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:52 crc kubenswrapper[4759]: I1125 19:51:52.974982 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:53 crc kubenswrapper[4759]: I1125 19:51:53.192044 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl"] Nov 25 19:51:53 crc kubenswrapper[4759]: I1125 19:51:53.654403 4759 generic.go:334] "Generic (PLEG): container finished" podID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerID="f113a486491991ab9eab8f095a3ad23f43dd3ad94fa31f747538edd9ada38803" exitCode=0 Nov 25 19:51:53 crc kubenswrapper[4759]: I1125 19:51:53.654486 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" event={"ID":"b4fc01b1-0131-4fa2-84bf-a4b000301ea8","Type":"ContainerDied","Data":"f113a486491991ab9eab8f095a3ad23f43dd3ad94fa31f747538edd9ada38803"} Nov 25 19:51:53 crc kubenswrapper[4759]: I1125 19:51:53.654601 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" event={"ID":"b4fc01b1-0131-4fa2-84bf-a4b000301ea8","Type":"ContainerStarted","Data":"6e019ec0519580409f8ce1d8ab7c16ce317ed7e2931428726289235814e0d8bb"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.688013 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"b86a8147c1fb920f9e92adb523e997ede4994fcba0614850e197d7a5db3a0fb6"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.689008 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"828d037417c3aa4d5674693f9cfc90bb485a1e8815a2ea2e586f7a883faeab5d"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.689023 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"3fdd22473668c8305c657dbf18db01ead478a7410c41060db88f84ff0444bb09"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.689035 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"dea318907fe0b45444842167cee695fd01cd11942817d8b2ffca1651148d082e"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.689044 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"7fd1b25338930477ca9c47492997c69eee436ba2bdf67b041f73836d2b968b2d"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.689052 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"ff8191ad91406493c21af5559caaa2a79598f6747d9beede956a4af3da8a98e2"} Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.691222 4759 generic.go:334] "Generic (PLEG): container finished" podID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerID="fc43cdfa31215da66aa465aea2f0fa73c77105d73de4eda8623b0bf2f5c3948b" exitCode=0 Nov 25 19:51:54 crc kubenswrapper[4759]: I1125 19:51:54.691270 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" event={"ID":"b4fc01b1-0131-4fa2-84bf-a4b000301ea8","Type":"ContainerDied","Data":"fc43cdfa31215da66aa465aea2f0fa73c77105d73de4eda8623b0bf2f5c3948b"} Nov 25 19:51:55 crc kubenswrapper[4759]: I1125 19:51:55.702707 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"07cbf89a-8a0b-48ab-9e23-52226f362fbc","Type":"ContainerStarted","Data":"a6d11dea0db19a53d6441c8ea902b026a24d7cf3f056dfaeb95d0b97cf0fcabd"} Nov 25 19:51:55 crc kubenswrapper[4759]: I1125 19:51:55.705531 4759 generic.go:334] "Generic (PLEG): container finished" podID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerID="3e1135972dfcbeccb25825efb243f7c9355450ddb054e2ddc97d7a7c8886c45b" exitCode=0 Nov 25 19:51:55 crc kubenswrapper[4759]: I1125 19:51:55.705558 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" event={"ID":"b4fc01b1-0131-4fa2-84bf-a4b000301ea8","Type":"ContainerDied","Data":"3e1135972dfcbeccb25825efb243f7c9355450ddb054e2ddc97d7a7c8886c45b"} Nov 25 19:51:55 crc kubenswrapper[4759]: I1125 19:51:55.747435 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=20.230279481 podStartE2EDuration="24.747420096s" podCreationTimestamp="2025-11-25 19:51:31 +0000 UTC" firstStartedPulling="2025-11-25 19:51:49.016737116 +0000 UTC m=+1099.602684817" lastFinishedPulling="2025-11-25 19:51:53.533877731 +0000 UTC m=+1104.119825432" observedRunningTime="2025-11-25 19:51:55.744993833 +0000 UTC m=+1106.330941534" watchObservedRunningTime="2025-11-25 19:51:55.747420096 +0000 UTC m=+1106.333367797" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.047723 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.203990 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle\") pod \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.204588 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util\") pod \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.204771 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2xvd\" (UniqueName: \"kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd\") pod \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\" (UID: \"b4fc01b1-0131-4fa2-84bf-a4b000301ea8\") " Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.206013 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle" (OuterVolumeSpecName: "bundle") pod "b4fc01b1-0131-4fa2-84bf-a4b000301ea8" (UID: "b4fc01b1-0131-4fa2-84bf-a4b000301ea8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.212652 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd" (OuterVolumeSpecName: "kube-api-access-b2xvd") pod "b4fc01b1-0131-4fa2-84bf-a4b000301ea8" (UID: "b4fc01b1-0131-4fa2-84bf-a4b000301ea8"). InnerVolumeSpecName "kube-api-access-b2xvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.219777 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util" (OuterVolumeSpecName: "util") pod "b4fc01b1-0131-4fa2-84bf-a4b000301ea8" (UID: "b4fc01b1-0131-4fa2-84bf-a4b000301ea8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.307288 4759 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.307332 4759 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-util\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.307346 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2xvd\" (UniqueName: \"kubernetes.io/projected/b4fc01b1-0131-4fa2-84bf-a4b000301ea8-kube-api-access-b2xvd\") on node \"crc\" DevicePath \"\"" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.722639 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" event={"ID":"b4fc01b1-0131-4fa2-84bf-a4b000301ea8","Type":"ContainerDied","Data":"6e019ec0519580409f8ce1d8ab7c16ce317ed7e2931428726289235814e0d8bb"} Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.722675 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e019ec0519580409f8ce1d8ab7c16ce317ed7e2931428726289235814e0d8bb" Nov 25 19:51:57 crc kubenswrapper[4759]: I1125 19:51:57.722734 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl" Nov 25 19:51:59 crc kubenswrapper[4759]: I1125 19:51:59.497635 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:51:59 crc kubenswrapper[4759]: I1125 19:51:59.503201 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.819326 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s"] Nov 25 19:52:19 crc kubenswrapper[4759]: E1125 19:52:19.819971 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="pull" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.819983 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="pull" Nov 25 19:52:19 crc kubenswrapper[4759]: E1125 19:52:19.820009 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="extract" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.820015 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="extract" Nov 25 19:52:19 crc kubenswrapper[4759]: E1125 19:52:19.820030 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="util" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.820037 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="util" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.820150 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4fc01b1-0131-4fa2-84bf-a4b000301ea8" containerName="extract" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.821666 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.824879 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.825030 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-69cn2" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.910152 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s"] Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.943202 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrm6z\" (UniqueName: \"kubernetes.io/projected/3843646a-ca2b-420e-b03b-a41a0e8e755a-kube-api-access-qrm6z\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.943277 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-apiservice-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:19 crc kubenswrapper[4759]: I1125 19:52:19.943383 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-webhook-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.044866 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-webhook-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.044921 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrm6z\" (UniqueName: \"kubernetes.io/projected/3843646a-ca2b-420e-b03b-a41a0e8e755a-kube-api-access-qrm6z\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.044956 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-apiservice-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.050680 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-apiservice-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.051935 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3843646a-ca2b-420e-b03b-a41a0e8e755a-webhook-cert\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.063190 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrm6z\" (UniqueName: \"kubernetes.io/projected/3843646a-ca2b-420e-b03b-a41a0e8e755a-kube-api-access-qrm6z\") pod \"glance-operator-controller-manager-85b98645d7-9h75s\" (UID: \"3843646a-ca2b-420e-b03b-a41a0e8e755a\") " pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.156767 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.574349 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s"] Nov 25 19:52:20 crc kubenswrapper[4759]: I1125 19:52:20.882900 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" event={"ID":"3843646a-ca2b-420e-b03b-a41a0e8e755a","Type":"ContainerStarted","Data":"f249aa8c46b8c1e35b416ad59b77f49a91aecf732d9b33b863d7c847dbdc2d62"} Nov 25 19:52:21 crc kubenswrapper[4759]: I1125 19:52:21.894232 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" event={"ID":"3843646a-ca2b-420e-b03b-a41a0e8e755a","Type":"ContainerStarted","Data":"3af8bc497d9d3729af7522076102b9a16a59f106047363564f64367315a53080"} Nov 25 19:52:21 crc kubenswrapper[4759]: I1125 19:52:21.894831 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:21 crc kubenswrapper[4759]: I1125 19:52:21.919904 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" podStartSLOduration=1.812376344 podStartE2EDuration="2.91987736s" podCreationTimestamp="2025-11-25 19:52:19 +0000 UTC" firstStartedPulling="2025-11-25 19:52:20.579538342 +0000 UTC m=+1131.165486043" lastFinishedPulling="2025-11-25 19:52:21.687039368 +0000 UTC m=+1132.272987059" observedRunningTime="2025-11-25 19:52:21.910666644 +0000 UTC m=+1132.496614385" watchObservedRunningTime="2025-11-25 19:52:21.91987736 +0000 UTC m=+1132.505825101" Nov 25 19:52:30 crc kubenswrapper[4759]: I1125 19:52:30.161726 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-85b98645d7-9h75s" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.682782 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-6g2z8"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.683738 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.686006 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.686951 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.688293 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.694869 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-6g2z8"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.704040 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.730083 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55gtj\" (UniqueName: \"kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.730141 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57qwb\" (UniqueName: \"kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.730175 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.730217 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.746830 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.748550 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.754538 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-95wfv" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.754667 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.755406 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.755465 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.762593 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832334 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832402 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832422 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832595 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55gtj\" (UniqueName: \"kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832650 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57qwb\" (UniqueName: \"kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832716 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832761 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n79b4\" (UniqueName: \"kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.832822 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.833551 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.833792 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.849841 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57qwb\" (UniqueName: \"kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb\") pod \"glance-5fe1-account-create-update-vxgv2\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.850672 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55gtj\" (UniqueName: \"kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj\") pod \"glance-db-create-6g2z8\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.934602 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.934698 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.934716 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.934765 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n79b4\" (UniqueName: \"kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.935674 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.935772 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.937584 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.953314 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n79b4\" (UniqueName: \"kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4\") pod \"openstackclient\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:31 crc kubenswrapper[4759]: I1125 19:52:31.999743 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.007153 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.083378 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 19:52:32 crc kubenswrapper[4759]: W1125 19:52:32.483602 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda3ff6ab_c18d_48cf_9f02_752df22a784e.slice/crio-aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4 WatchSource:0}: Error finding container aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4: Status 404 returned error can't find the container with id aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4 Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.484358 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2"] Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.516609 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-6g2z8"] Nov 25 19:52:32 crc kubenswrapper[4759]: W1125 19:52:32.523539 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46135ac8_9252_4497_8c18_480c1e0c9157.slice/crio-2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0 WatchSource:0}: Error finding container 2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0: Status 404 returned error can't find the container with id 2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0 Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.564759 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.974500 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6","Type":"ContainerStarted","Data":"ff3b1c71d34e91fafc67405f7fefdc2a5c342db587648fa58c016fa16c84a965"} Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.976618 4759 generic.go:334] "Generic (PLEG): container finished" podID="46135ac8-9252-4497-8c18-480c1e0c9157" containerID="ff55e90629e22372aebdf02d0242a6aee9ea7beda4c51ac0654b12807c33b920" exitCode=0 Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.976674 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6g2z8" event={"ID":"46135ac8-9252-4497-8c18-480c1e0c9157","Type":"ContainerDied","Data":"ff55e90629e22372aebdf02d0242a6aee9ea7beda4c51ac0654b12807c33b920"} Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.976695 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6g2z8" event={"ID":"46135ac8-9252-4497-8c18-480c1e0c9157","Type":"ContainerStarted","Data":"2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0"} Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.978196 4759 generic.go:334] "Generic (PLEG): container finished" podID="da3ff6ab-c18d-48cf-9f02-752df22a784e" containerID="e37d654761b2b5e0ae774004dcd1fee6685d57e31e272fced6511fe3a5de485c" exitCode=0 Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.978231 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" event={"ID":"da3ff6ab-c18d-48cf-9f02-752df22a784e","Type":"ContainerDied","Data":"e37d654761b2b5e0ae774004dcd1fee6685d57e31e272fced6511fe3a5de485c"} Nov 25 19:52:32 crc kubenswrapper[4759]: I1125 19:52:32.978249 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" event={"ID":"da3ff6ab-c18d-48cf-9f02-752df22a784e","Type":"ContainerStarted","Data":"aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4"} Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.440791 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.446429 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.603466 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55gtj\" (UniqueName: \"kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj\") pod \"46135ac8-9252-4497-8c18-480c1e0c9157\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.603530 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57qwb\" (UniqueName: \"kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb\") pod \"da3ff6ab-c18d-48cf-9f02-752df22a784e\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.603577 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts\") pod \"da3ff6ab-c18d-48cf-9f02-752df22a784e\" (UID: \"da3ff6ab-c18d-48cf-9f02-752df22a784e\") " Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.603599 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts\") pod \"46135ac8-9252-4497-8c18-480c1e0c9157\" (UID: \"46135ac8-9252-4497-8c18-480c1e0c9157\") " Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.604809 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da3ff6ab-c18d-48cf-9f02-752df22a784e" (UID: "da3ff6ab-c18d-48cf-9f02-752df22a784e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.604953 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "46135ac8-9252-4497-8c18-480c1e0c9157" (UID: "46135ac8-9252-4497-8c18-480c1e0c9157"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.608476 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj" (OuterVolumeSpecName: "kube-api-access-55gtj") pod "46135ac8-9252-4497-8c18-480c1e0c9157" (UID: "46135ac8-9252-4497-8c18-480c1e0c9157"). InnerVolumeSpecName "kube-api-access-55gtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.610095 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb" (OuterVolumeSpecName: "kube-api-access-57qwb") pod "da3ff6ab-c18d-48cf-9f02-752df22a784e" (UID: "da3ff6ab-c18d-48cf-9f02-752df22a784e"). InnerVolumeSpecName "kube-api-access-57qwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.704817 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da3ff6ab-c18d-48cf-9f02-752df22a784e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.704861 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46135ac8-9252-4497-8c18-480c1e0c9157-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.704874 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55gtj\" (UniqueName: \"kubernetes.io/projected/46135ac8-9252-4497-8c18-480c1e0c9157-kube-api-access-55gtj\") on node \"crc\" DevicePath \"\"" Nov 25 19:52:34 crc kubenswrapper[4759]: I1125 19:52:34.704889 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57qwb\" (UniqueName: \"kubernetes.io/projected/da3ff6ab-c18d-48cf-9f02-752df22a784e-kube-api-access-57qwb\") on node \"crc\" DevicePath \"\"" Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.015744 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6g2z8" event={"ID":"46135ac8-9252-4497-8c18-480c1e0c9157","Type":"ContainerDied","Data":"2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0"} Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.016014 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2983211b1cc29d263317aff323d2c644b8e5ea15ba4aad1f246d2733ea120ff0" Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.015773 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6g2z8" Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.017484 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" event={"ID":"da3ff6ab-c18d-48cf-9f02-752df22a784e","Type":"ContainerDied","Data":"aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4"} Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.017507 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2" Nov 25 19:52:35 crc kubenswrapper[4759]: I1125 19:52:35.017556 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aac44762d47cc1a0d75dd1979393dda287197dd01b4b88137f3916a82bb83ed4" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.827546 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-d7jb2"] Nov 25 19:52:36 crc kubenswrapper[4759]: E1125 19:52:36.827913 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46135ac8-9252-4497-8c18-480c1e0c9157" containerName="mariadb-database-create" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.827930 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="46135ac8-9252-4497-8c18-480c1e0c9157" containerName="mariadb-database-create" Nov 25 19:52:36 crc kubenswrapper[4759]: E1125 19:52:36.827947 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3ff6ab-c18d-48cf-9f02-752df22a784e" containerName="mariadb-account-create-update" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.827955 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3ff6ab-c18d-48cf-9f02-752df22a784e" containerName="mariadb-account-create-update" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.828131 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="46135ac8-9252-4497-8c18-480c1e0c9157" containerName="mariadb-database-create" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.828153 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3ff6ab-c18d-48cf-9f02-752df22a784e" containerName="mariadb-account-create-update" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.828759 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.832373 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.832913 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-j7qhq" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.834588 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d7jb2"] Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.945430 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.946004 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwqvv\" (UniqueName: \"kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:36 crc kubenswrapper[4759]: I1125 19:52:36.946124 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.047789 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.048050 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwqvv\" (UniqueName: \"kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.048138 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.056034 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.056086 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.063757 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwqvv\" (UniqueName: \"kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv\") pod \"glance-db-sync-d7jb2\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:37 crc kubenswrapper[4759]: I1125 19:52:37.149703 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:52:41 crc kubenswrapper[4759]: I1125 19:52:41.022910 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d7jb2"] Nov 25 19:52:41 crc kubenswrapper[4759]: W1125 19:52:41.027085 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe809738_f412_4899_9637_cb0c8fb43aaa.slice/crio-1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b WatchSource:0}: Error finding container 1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b: Status 404 returned error can't find the container with id 1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b Nov 25 19:52:41 crc kubenswrapper[4759]: I1125 19:52:41.057469 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6","Type":"ContainerStarted","Data":"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c"} Nov 25 19:52:41 crc kubenswrapper[4759]: I1125 19:52:41.058355 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d7jb2" event={"ID":"be809738-f412-4899-9637-cb0c8fb43aaa","Type":"ContainerStarted","Data":"1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b"} Nov 25 19:52:41 crc kubenswrapper[4759]: I1125 19:52:41.077827 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.980616993 podStartE2EDuration="10.077802018s" podCreationTimestamp="2025-11-25 19:52:31 +0000 UTC" firstStartedPulling="2025-11-25 19:52:32.570668212 +0000 UTC m=+1143.156615913" lastFinishedPulling="2025-11-25 19:52:40.667853237 +0000 UTC m=+1151.253800938" observedRunningTime="2025-11-25 19:52:41.070938672 +0000 UTC m=+1151.656886373" watchObservedRunningTime="2025-11-25 19:52:41.077802018 +0000 UTC m=+1151.663749759" Nov 25 19:52:53 crc kubenswrapper[4759]: I1125 19:52:53.166076 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d7jb2" event={"ID":"be809738-f412-4899-9637-cb0c8fb43aaa","Type":"ContainerStarted","Data":"6f6ce9372ba72e277d6ff0bd00522e4e3f755795987065d119db9f479d322af6"} Nov 25 19:52:53 crc kubenswrapper[4759]: I1125 19:52:53.185604 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-d7jb2" podStartSLOduration=6.496065694 podStartE2EDuration="17.185590252s" podCreationTimestamp="2025-11-25 19:52:36 +0000 UTC" firstStartedPulling="2025-11-25 19:52:41.029072034 +0000 UTC m=+1151.615019765" lastFinishedPulling="2025-11-25 19:52:51.718596622 +0000 UTC m=+1162.304544323" observedRunningTime="2025-11-25 19:52:53.181694992 +0000 UTC m=+1163.767642693" watchObservedRunningTime="2025-11-25 19:52:53.185590252 +0000 UTC m=+1163.771537953" Nov 25 19:53:00 crc kubenswrapper[4759]: I1125 19:53:00.252921 4759 generic.go:334] "Generic (PLEG): container finished" podID="be809738-f412-4899-9637-cb0c8fb43aaa" containerID="6f6ce9372ba72e277d6ff0bd00522e4e3f755795987065d119db9f479d322af6" exitCode=0 Nov 25 19:53:00 crc kubenswrapper[4759]: I1125 19:53:00.253057 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d7jb2" event={"ID":"be809738-f412-4899-9637-cb0c8fb43aaa","Type":"ContainerDied","Data":"6f6ce9372ba72e277d6ff0bd00522e4e3f755795987065d119db9f479d322af6"} Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.628664 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.710487 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data\") pod \"be809738-f412-4899-9637-cb0c8fb43aaa\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.710608 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwqvv\" (UniqueName: \"kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv\") pod \"be809738-f412-4899-9637-cb0c8fb43aaa\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.710645 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data\") pod \"be809738-f412-4899-9637-cb0c8fb43aaa\" (UID: \"be809738-f412-4899-9637-cb0c8fb43aaa\") " Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.716707 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "be809738-f412-4899-9637-cb0c8fb43aaa" (UID: "be809738-f412-4899-9637-cb0c8fb43aaa"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.716765 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv" (OuterVolumeSpecName: "kube-api-access-dwqvv") pod "be809738-f412-4899-9637-cb0c8fb43aaa" (UID: "be809738-f412-4899-9637-cb0c8fb43aaa"). InnerVolumeSpecName "kube-api-access-dwqvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.747748 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data" (OuterVolumeSpecName: "config-data") pod "be809738-f412-4899-9637-cb0c8fb43aaa" (UID: "be809738-f412-4899-9637-cb0c8fb43aaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.812459 4759 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.812495 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwqvv\" (UniqueName: \"kubernetes.io/projected/be809738-f412-4899-9637-cb0c8fb43aaa-kube-api-access-dwqvv\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:01 crc kubenswrapper[4759]: I1125 19:53:01.812512 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be809738-f412-4899-9637-cb0c8fb43aaa-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.272256 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d7jb2" event={"ID":"be809738-f412-4899-9637-cb0c8fb43aaa","Type":"ContainerDied","Data":"1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b"} Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.272311 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1757caed067937952bf88b40df918cf95284ed0dccd33739049cb5abf315480b" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.272355 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d7jb2" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.945181 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:02 crc kubenswrapper[4759]: E1125 19:53:02.945680 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be809738-f412-4899-9637-cb0c8fb43aaa" containerName="glance-db-sync" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.945691 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="be809738-f412-4899-9637-cb0c8fb43aaa" containerName="glance-db-sync" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.945820 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="be809738-f412-4899-9637-cb0c8fb43aaa" containerName="glance-db-sync" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.946477 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.948181 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-j7qhq" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.948598 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.952143 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.961745 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.969879 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.970944 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:02 crc kubenswrapper[4759]: I1125 19:53:02.990218 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029430 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029526 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029578 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029599 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029632 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dptgk\" (UniqueName: \"kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029650 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029724 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029757 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029787 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029815 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029834 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029851 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029874 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.029901 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131080 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131139 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131175 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g85sf\" (UniqueName: \"kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131208 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131243 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131274 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131296 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131331 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131360 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131396 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131473 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131510 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131533 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131574 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131580 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131608 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131638 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131670 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131696 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dptgk\" (UniqueName: \"kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131721 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131749 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131777 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131806 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131828 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131849 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131880 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131911 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131933 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131957 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.131978 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132051 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132078 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132201 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132307 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132346 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132382 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132404 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.132727 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.134317 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.136271 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.139959 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.151525 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dptgk\" (UniqueName: \"kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.157313 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.183043 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233206 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233247 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233301 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233348 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233372 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233400 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233416 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233431 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233464 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233491 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233513 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233541 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g85sf\" (UniqueName: \"kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233566 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.233584 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.234172 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.234242 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235015 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235125 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235187 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235212 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235249 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235349 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235273 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235389 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.235511 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.238236 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.240223 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.255677 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.255966 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.260029 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g85sf\" (UniqueName: \"kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf\") pod \"glance-default-single-1\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.260374 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.285687 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.715976 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:03 crc kubenswrapper[4759]: I1125 19:53:03.782513 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:03 crc kubenswrapper[4759]: W1125 19:53:03.786978 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4089934c_02b9_408e_b674_0ccd2dbd84cf.slice/crio-5063e0c62b16c9e5e1c10b5766ef86430fd7b5999accd344bb46b512924a25dc WatchSource:0}: Error finding container 5063e0c62b16c9e5e1c10b5766ef86430fd7b5999accd344bb46b512924a25dc: Status 404 returned error can't find the container with id 5063e0c62b16c9e5e1c10b5766ef86430fd7b5999accd344bb46b512924a25dc Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.070168 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.286894 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerStarted","Data":"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.287215 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerStarted","Data":"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.287228 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerStarted","Data":"5063e0c62b16c9e5e1c10b5766ef86430fd7b5999accd344bb46b512924a25dc"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.287025 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-httpd" containerID="cri-o://c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" gracePeriod=30 Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.286986 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-log" containerID="cri-o://a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" gracePeriod=30 Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.292948 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerStarted","Data":"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.292992 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerStarted","Data":"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.293005 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerStarted","Data":"682d3719c013c1cc15c6284c0cedb2034f685780df4d86069b4df06f51e5ce45"} Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.309301 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.309281326 podStartE2EDuration="2.309281326s" podCreationTimestamp="2025-11-25 19:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:53:04.30517118 +0000 UTC m=+1174.891118881" watchObservedRunningTime="2025-11-25 19:53:04.309281326 +0000 UTC m=+1174.895229027" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.328476 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.328435639 podStartE2EDuration="2.328435639s" podCreationTimestamp="2025-11-25 19:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:53:04.325144555 +0000 UTC m=+1174.911092266" watchObservedRunningTime="2025-11-25 19:53:04.328435639 +0000 UTC m=+1174.914383340" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.651012 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755777 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755841 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755873 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755902 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755939 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755981 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.755995 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756011 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756040 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756096 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756115 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g85sf\" (UniqueName: \"kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756141 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756160 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756181 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data\") pod \"4089934c-02b9-408e-b674-0ccd2dbd84cf\" (UID: \"4089934c-02b9-408e-b674-0ccd2dbd84cf\") " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.756517 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757184 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys" (OuterVolumeSpecName: "sys") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757251 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev" (OuterVolumeSpecName: "dev") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757286 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757263 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757268 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs" (OuterVolumeSpecName: "logs") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757288 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run" (OuterVolumeSpecName: "run") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757621 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.757694 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.762143 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.762268 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts" (OuterVolumeSpecName: "scripts") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.762370 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.762375 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf" (OuterVolumeSpecName: "kube-api-access-g85sf") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "kube-api-access-g85sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.809865 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data" (OuterVolumeSpecName: "config-data") pod "4089934c-02b9-408e-b674-0ccd2dbd84cf" (UID: "4089934c-02b9-408e-b674-0ccd2dbd84cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.858890 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.858941 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.858954 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859003 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859019 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859041 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859054 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859066 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859080 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g85sf\" (UniqueName: \"kubernetes.io/projected/4089934c-02b9-408e-b674-0ccd2dbd84cf-kube-api-access-g85sf\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859094 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859105 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859116 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4089934c-02b9-408e-b674-0ccd2dbd84cf-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859127 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4089934c-02b9-408e-b674-0ccd2dbd84cf-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.859138 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4089934c-02b9-408e-b674-0ccd2dbd84cf-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.877349 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.886373 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.960577 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:04 crc kubenswrapper[4759]: I1125 19:53:04.960609 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.299665 4759 generic.go:334] "Generic (PLEG): container finished" podID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerID="c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" exitCode=143 Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.299994 4759 generic.go:334] "Generic (PLEG): container finished" podID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerID="a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" exitCode=143 Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.299756 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.299771 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerDied","Data":"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a"} Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.300060 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerDied","Data":"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d"} Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.300079 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4089934c-02b9-408e-b674-0ccd2dbd84cf","Type":"ContainerDied","Data":"5063e0c62b16c9e5e1c10b5766ef86430fd7b5999accd344bb46b512924a25dc"} Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.300098 4759 scope.go:117] "RemoveContainer" containerID="c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.329658 4759 scope.go:117] "RemoveContainer" containerID="a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.331361 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.339212 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.355087 4759 scope.go:117] "RemoveContainer" containerID="c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" Nov 25 19:53:05 crc kubenswrapper[4759]: E1125 19:53:05.357013 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a\": container with ID starting with c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a not found: ID does not exist" containerID="c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.357057 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a"} err="failed to get container status \"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a\": rpc error: code = NotFound desc = could not find container \"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a\": container with ID starting with c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a not found: ID does not exist" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.357086 4759 scope.go:117] "RemoveContainer" containerID="a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" Nov 25 19:53:05 crc kubenswrapper[4759]: E1125 19:53:05.357576 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d\": container with ID starting with a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d not found: ID does not exist" containerID="a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.357604 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d"} err="failed to get container status \"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d\": rpc error: code = NotFound desc = could not find container \"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d\": container with ID starting with a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d not found: ID does not exist" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.357622 4759 scope.go:117] "RemoveContainer" containerID="c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.359075 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a"} err="failed to get container status \"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a\": rpc error: code = NotFound desc = could not find container \"c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a\": container with ID starting with c41072cdda79ec8cba0ca8afa46bcfa511814bdd3a398193500dbf3e90ea8c7a not found: ID does not exist" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.359125 4759 scope.go:117] "RemoveContainer" containerID="a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.361668 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d"} err="failed to get container status \"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d\": rpc error: code = NotFound desc = could not find container \"a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d\": container with ID starting with a894399b43e56497c9f4e8817329c0f8a8e50596f97133aab8e1aa017cdd3a6d not found: ID does not exist" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.362914 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:05 crc kubenswrapper[4759]: E1125 19:53:05.363243 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-log" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.363302 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-log" Nov 25 19:53:05 crc kubenswrapper[4759]: E1125 19:53:05.363334 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-httpd" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.363344 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-httpd" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.363553 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-log" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.363581 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" containerName="glance-httpd" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.364420 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.397068 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568147 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568198 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568229 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568303 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568330 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568359 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568385 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568404 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568441 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568523 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568580 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568637 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568658 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.568678 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgvh\" (UniqueName: \"kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.671860 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.671907 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.671934 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgvh\" (UniqueName: \"kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.671969 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.671989 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672013 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672013 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672050 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672077 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672121 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672152 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672173 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672211 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672254 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672292 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672498 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672120 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672587 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.672834 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673046 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673114 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673417 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673803 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673811 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.673870 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.680260 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.682542 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.705992 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgvh\" (UniqueName: \"kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.710382 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.711223 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:05 crc kubenswrapper[4759]: I1125 19:53:05.997108 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:06 crc kubenswrapper[4759]: I1125 19:53:06.115979 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4089934c-02b9-408e-b674-0ccd2dbd84cf" path="/var/lib/kubelet/pods/4089934c-02b9-408e-b674-0ccd2dbd84cf/volumes" Nov 25 19:53:06 crc kubenswrapper[4759]: I1125 19:53:06.407047 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:07 crc kubenswrapper[4759]: I1125 19:53:07.317754 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerStarted","Data":"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553"} Nov 25 19:53:07 crc kubenswrapper[4759]: I1125 19:53:07.318667 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerStarted","Data":"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0"} Nov 25 19:53:07 crc kubenswrapper[4759]: I1125 19:53:07.318703 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerStarted","Data":"98add4845fcda33a45dc700d5fd4ee34d1490812b9a1433f955c3aeca9b4198a"} Nov 25 19:53:07 crc kubenswrapper[4759]: I1125 19:53:07.359803 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.359781263 podStartE2EDuration="2.359781263s" podCreationTimestamp="2025-11-25 19:53:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:53:07.35190706 +0000 UTC m=+1177.937854811" watchObservedRunningTime="2025-11-25 19:53:07.359781263 +0000 UTC m=+1177.945728974" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.261309 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.262909 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.290198 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.315873 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.364084 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:13 crc kubenswrapper[4759]: I1125 19:53:13.364128 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:15 crc kubenswrapper[4759]: I1125 19:53:15.299740 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:15 crc kubenswrapper[4759]: I1125 19:53:15.349221 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:15 crc kubenswrapper[4759]: I1125 19:53:15.998164 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:15 crc kubenswrapper[4759]: I1125 19:53:15.998284 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:16 crc kubenswrapper[4759]: I1125 19:53:16.038164 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:16 crc kubenswrapper[4759]: I1125 19:53:16.041243 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:16 crc kubenswrapper[4759]: I1125 19:53:16.385012 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:16 crc kubenswrapper[4759]: I1125 19:53:16.385057 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:18 crc kubenswrapper[4759]: I1125 19:53:18.317095 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:18 crc kubenswrapper[4759]: I1125 19:53:18.318897 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:18 crc kubenswrapper[4759]: I1125 19:53:18.401859 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:18 crc kubenswrapper[4759]: I1125 19:53:18.402083 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-log" containerID="cri-o://dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32" gracePeriod=30 Nov 25 19:53:18 crc kubenswrapper[4759]: I1125 19:53:18.402290 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-httpd" containerID="cri-o://10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f" gracePeriod=30 Nov 25 19:53:19 crc kubenswrapper[4759]: I1125 19:53:19.412520 4759 generic.go:334] "Generic (PLEG): container finished" podID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerID="dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32" exitCode=143 Nov 25 19:53:19 crc kubenswrapper[4759]: I1125 19:53:19.412603 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerDied","Data":"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32"} Nov 25 19:53:21 crc kubenswrapper[4759]: I1125 19:53:21.987460 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117091 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117122 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117146 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117182 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117218 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117236 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117287 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117303 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117328 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117342 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117337 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117368 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117384 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117386 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys" (OuterVolumeSpecName: "sys") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117405 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117426 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117461 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dptgk\" (UniqueName: \"kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk\") pod \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\" (UID: \"9ab1081b-04c8-4142-ad32-36dd3a0e828a\") " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117496 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117546 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117567 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117711 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117723 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117733 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117741 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117749 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117757 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117817 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs" (OuterVolumeSpecName: "logs") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117850 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev" (OuterVolumeSpecName: "dev") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.117862 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run" (OuterVolumeSpecName: "run") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.123141 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.126982 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.127993 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts" (OuterVolumeSpecName: "scripts") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.137539 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk" (OuterVolumeSpecName: "kube-api-access-dptgk") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "kube-api-access-dptgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.156833 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data" (OuterVolumeSpecName: "config-data") pod "9ab1081b-04c8-4142-ad32-36dd3a0e828a" (UID: "9ab1081b-04c8-4142-ad32-36dd3a0e828a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219081 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab1081b-04c8-4142-ad32-36dd3a0e828a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219352 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219433 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dptgk\" (UniqueName: \"kubernetes.io/projected/9ab1081b-04c8-4142-ad32-36dd3a0e828a-kube-api-access-dptgk\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219521 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219578 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab1081b-04c8-4142-ad32-36dd3a0e828a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219643 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ab1081b-04c8-4142-ad32-36dd3a0e828a-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219719 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.219780 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.234737 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.236683 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.321136 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.321171 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.436543 4759 generic.go:334] "Generic (PLEG): container finished" podID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerID="10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f" exitCode=0 Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.436582 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerDied","Data":"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f"} Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.436607 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9ab1081b-04c8-4142-ad32-36dd3a0e828a","Type":"ContainerDied","Data":"682d3719c013c1cc15c6284c0cedb2034f685780df4d86069b4df06f51e5ce45"} Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.436623 4759 scope.go:117] "RemoveContainer" containerID="10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.436733 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.470110 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.479104 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.480425 4759 scope.go:117] "RemoveContainer" containerID="dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.499900 4759 scope.go:117] "RemoveContainer" containerID="10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f" Nov 25 19:53:22 crc kubenswrapper[4759]: E1125 19:53:22.500424 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f\": container with ID starting with 10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f not found: ID does not exist" containerID="10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.500475 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f"} err="failed to get container status \"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f\": rpc error: code = NotFound desc = could not find container \"10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f\": container with ID starting with 10de129c64c0bf55d7f3d60ede5e3abe6f2b5ebe5ca7931285e367725cb7700f not found: ID does not exist" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.500498 4759 scope.go:117] "RemoveContainer" containerID="dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32" Nov 25 19:53:22 crc kubenswrapper[4759]: E1125 19:53:22.500835 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32\": container with ID starting with dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32 not found: ID does not exist" containerID="dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.501049 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32"} err="failed to get container status \"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32\": rpc error: code = NotFound desc = could not find container \"dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32\": container with ID starting with dcc4904abfcfee1a1090acccdd517c18feba21bc6c13618d1c9d1105affebb32 not found: ID does not exist" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.518886 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:22 crc kubenswrapper[4759]: E1125 19:53:22.519223 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-log" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.519240 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-log" Nov 25 19:53:22 crc kubenswrapper[4759]: E1125 19:53:22.519256 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-httpd" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.519264 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-httpd" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.519476 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-log" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.519498 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" containerName="glance-httpd" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.521216 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.530376 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625390 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625434 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625481 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625517 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhwn\" (UniqueName: \"kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625593 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625622 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625699 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625720 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625736 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625797 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625904 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625951 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625970 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.625986 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727515 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727597 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727671 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727692 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727671 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727934 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727968 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhwn\" (UniqueName: \"kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.727974 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728001 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728027 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728070 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728231 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728260 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728282 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728284 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728257 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728333 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728363 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728411 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728433 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728494 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728548 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.728572 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.729043 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.729223 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.733821 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.743258 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.746711 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhwn\" (UniqueName: \"kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.771293 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.793165 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:22 crc kubenswrapper[4759]: I1125 19:53:22.841513 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:23 crc kubenswrapper[4759]: I1125 19:53:23.266080 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:23 crc kubenswrapper[4759]: I1125 19:53:23.448912 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerStarted","Data":"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f"} Nov 25 19:53:23 crc kubenswrapper[4759]: I1125 19:53:23.449271 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerStarted","Data":"f9632e81ba3fbab710a1ab05cffdfc07cf5fbd3a783cdb7baf4fe5b54247b376"} Nov 25 19:53:24 crc kubenswrapper[4759]: I1125 19:53:24.119871 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab1081b-04c8-4142-ad32-36dd3a0e828a" path="/var/lib/kubelet/pods/9ab1081b-04c8-4142-ad32-36dd3a0e828a/volumes" Nov 25 19:53:24 crc kubenswrapper[4759]: I1125 19:53:24.459675 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerStarted","Data":"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a"} Nov 25 19:53:24 crc kubenswrapper[4759]: I1125 19:53:24.510762 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.510742542 podStartE2EDuration="2.510742542s" podCreationTimestamp="2025-11-25 19:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:53:24.502610913 +0000 UTC m=+1195.088558644" watchObservedRunningTime="2025-11-25 19:53:24.510742542 +0000 UTC m=+1195.096690253" Nov 25 19:53:24 crc kubenswrapper[4759]: I1125 19:53:24.895664 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:53:24 crc kubenswrapper[4759]: I1125 19:53:24.895970 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:53:32 crc kubenswrapper[4759]: I1125 19:53:32.843101 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:32 crc kubenswrapper[4759]: I1125 19:53:32.843656 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:32 crc kubenswrapper[4759]: I1125 19:53:32.866724 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:32 crc kubenswrapper[4759]: I1125 19:53:32.879766 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:33 crc kubenswrapper[4759]: I1125 19:53:33.524055 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:33 crc kubenswrapper[4759]: I1125 19:53:33.524090 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:35 crc kubenswrapper[4759]: I1125 19:53:35.392997 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:35 crc kubenswrapper[4759]: I1125 19:53:35.414001 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.443098 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d7jb2"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.448278 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d7jb2"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.501726 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.501963 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-log" containerID="cri-o://b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0" gracePeriod=30 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.502035 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-httpd" containerID="cri-o://6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553" gracePeriod=30 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.524984 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.525254 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-log" containerID="cri-o://8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f" gracePeriod=30 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.525398 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-httpd" containerID="cri-o://d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a" gracePeriod=30 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.578743 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance5fe1-account-delete-bp2nk"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.589576 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.599916 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance5fe1-account-delete-bp2nk"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.655254 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.655594 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" containerName="openstackclient" containerID="cri-o://12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c" gracePeriod=30 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.666048 4759 generic.go:334] "Generic (PLEG): container finished" podID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerID="b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0" exitCode=143 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.666133 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerDied","Data":"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0"} Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.668098 4759 generic.go:334] "Generic (PLEG): container finished" podID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerID="8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f" exitCode=143 Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.668130 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerDied","Data":"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f"} Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.747324 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsrjb\" (UniqueName: \"kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.747554 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.849933 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.850735 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrjb\" (UniqueName: \"kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.850921 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.871976 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrjb\" (UniqueName: \"kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb\") pod \"glance5fe1-account-delete-bp2nk\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:51 crc kubenswrapper[4759]: I1125 19:53:51.913733 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.105561 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.117227 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be809738-f412-4899-9637-cb0c8fb43aaa" path="/var/lib/kubelet/pods/be809738-f412-4899-9637-cb0c8fb43aaa/volumes" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.256472 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n79b4\" (UniqueName: \"kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4\") pod \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.256564 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts\") pod \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.256667 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret\") pod \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.256727 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config\") pod \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\" (UID: \"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6\") " Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.257400 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" (UID: "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.264147 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4" (OuterVolumeSpecName: "kube-api-access-n79b4") pod "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" (UID: "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6"). InnerVolumeSpecName "kube-api-access-n79b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.273830 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" (UID: "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.278087 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" (UID: "bcbe3a07-6001-4b37-bb3f-bc31e30fcce6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.358459 4759 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.358494 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n79b4\" (UniqueName: \"kubernetes.io/projected/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-kube-api-access-n79b4\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.358508 4759 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.358520 4759 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.373792 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance5fe1-account-delete-bp2nk"] Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.678349 4759 generic.go:334] "Generic (PLEG): container finished" podID="759d2f0f-0f92-4d86-b3a2-09301802f2a8" containerID="c051c659087c6b9bcb1fbd7a9b36f64920c5039e1b9b595eb0fa4c47c400cc97" exitCode=0 Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.678480 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" event={"ID":"759d2f0f-0f92-4d86-b3a2-09301802f2a8","Type":"ContainerDied","Data":"c051c659087c6b9bcb1fbd7a9b36f64920c5039e1b9b595eb0fa4c47c400cc97"} Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.679162 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" event={"ID":"759d2f0f-0f92-4d86-b3a2-09301802f2a8","Type":"ContainerStarted","Data":"73f21db8ebf553ac8c804325badd000a9447ac3936a03ba1969cbe065dd2c07b"} Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.681439 4759 generic.go:334] "Generic (PLEG): container finished" podID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" containerID="12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c" exitCode=143 Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.681508 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6","Type":"ContainerDied","Data":"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c"} Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.681546 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"bcbe3a07-6001-4b37-bb3f-bc31e30fcce6","Type":"ContainerDied","Data":"ff3b1c71d34e91fafc67405f7fefdc2a5c342db587648fa58c016fa16c84a965"} Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.681555 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.681592 4759 scope.go:117] "RemoveContainer" containerID="12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.724918 4759 scope.go:117] "RemoveContainer" containerID="12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c" Nov 25 19:53:52 crc kubenswrapper[4759]: E1125 19:53:52.725413 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c\": container with ID starting with 12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c not found: ID does not exist" containerID="12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.725461 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c"} err="failed to get container status \"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c\": rpc error: code = NotFound desc = could not find container \"12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c\": container with ID starting with 12761443117f36118f32a2b703c94989cea47eb61c17d709d25a08c1901c799c not found: ID does not exist" Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.739531 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:53:52 crc kubenswrapper[4759]: I1125 19:53:52.744903 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 19:53:53 crc kubenswrapper[4759]: I1125 19:53:53.946131 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.084652 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsrjb\" (UniqueName: \"kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb\") pod \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.084705 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts\") pod \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\" (UID: \"759d2f0f-0f92-4d86-b3a2-09301802f2a8\") " Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.085591 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "759d2f0f-0f92-4d86-b3a2-09301802f2a8" (UID: "759d2f0f-0f92-4d86-b3a2-09301802f2a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.089485 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb" (OuterVolumeSpecName: "kube-api-access-rsrjb") pod "759d2f0f-0f92-4d86-b3a2-09301802f2a8" (UID: "759d2f0f-0f92-4d86-b3a2-09301802f2a8"). InnerVolumeSpecName "kube-api-access-rsrjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.118354 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" path="/var/lib/kubelet/pods/bcbe3a07-6001-4b37-bb3f-bc31e30fcce6/volumes" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.187154 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsrjb\" (UniqueName: \"kubernetes.io/projected/759d2f0f-0f92-4d86-b3a2-09301802f2a8-kube-api-access-rsrjb\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.187230 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759d2f0f-0f92-4d86-b3a2-09301802f2a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.697219 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.102:9292/healthcheck\": read tcp 10.217.0.2:45512->10.217.0.102:9292: read: connection reset by peer" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.697269 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.102:9292/healthcheck\": read tcp 10.217.0.2:45498->10.217.0.102:9292: read: connection reset by peer" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.700020 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" event={"ID":"759d2f0f-0f92-4d86-b3a2-09301802f2a8","Type":"ContainerDied","Data":"73f21db8ebf553ac8c804325badd000a9447ac3936a03ba1969cbe065dd2c07b"} Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.700053 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f21db8ebf553ac8c804325badd000a9447ac3936a03ba1969cbe065dd2c07b" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.700149 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5fe1-account-delete-bp2nk" Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.895791 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:53:54 crc kubenswrapper[4759]: I1125 19:53:54.896112 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.065635 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209533 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209590 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209638 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxhwn\" (UniqueName: \"kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209672 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209708 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209737 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209771 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209808 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209856 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209910 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209929 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209961 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.209995 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.210023 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run\") pod \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\" (UID: \"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213478 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213497 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213522 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys" (OuterVolumeSpecName: "sys") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213537 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213545 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213665 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run" (OuterVolumeSpecName: "run") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213697 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev" (OuterVolumeSpecName: "dev") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.213901 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs" (OuterVolumeSpecName: "logs") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.214304 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.216576 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.217173 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn" (OuterVolumeSpecName: "kube-api-access-gxhwn") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "kube-api-access-gxhwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.217583 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.218278 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts" (OuterVolumeSpecName: "scripts") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.267601 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data" (OuterVolumeSpecName: "config-data") pod "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" (UID: "aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.302843 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311627 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311657 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311668 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311676 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311684 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311692 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311721 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311733 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311743 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxhwn\" (UniqueName: \"kubernetes.io/projected/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-kube-api-access-gxhwn\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311754 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311762 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311770 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311778 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.311788 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.327365 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.329650 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412327 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412362 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412387 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412412 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412430 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412483 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412488 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412534 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412570 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srgvh\" (UniqueName: \"kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412587 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412603 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412632 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412656 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412738 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412754 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys\") pod \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\" (UID: \"32d1d95a-1133-49b7-a59a-5a7687e4de2f\") " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.412779 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413000 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413011 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413018 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413026 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413046 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys" (OuterVolumeSpecName: "sys") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413285 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev" (OuterVolumeSpecName: "dev") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413298 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413342 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413356 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run" (OuterVolumeSpecName: "run") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413372 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.413598 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs" (OuterVolumeSpecName: "logs") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.415489 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.415502 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh" (OuterVolumeSpecName: "kube-api-access-srgvh") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "kube-api-access-srgvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.416272 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.416376 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts" (OuterVolumeSpecName: "scripts") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.442880 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data" (OuterVolumeSpecName: "config-data") pod "32d1d95a-1133-49b7-a59a-5a7687e4de2f" (UID: "32d1d95a-1133-49b7-a59a-5a7687e4de2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514129 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514164 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514200 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514215 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514225 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514235 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srgvh\" (UniqueName: \"kubernetes.io/projected/32d1d95a-1133-49b7-a59a-5a7687e4de2f-kube-api-access-srgvh\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514248 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d1d95a-1133-49b7-a59a-5a7687e4de2f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514257 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514267 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514276 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d1d95a-1133-49b7-a59a-5a7687e4de2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514286 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.514294 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/32d1d95a-1133-49b7-a59a-5a7687e4de2f-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.525651 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.526145 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.615459 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.615493 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.708263 4759 generic.go:334] "Generic (PLEG): container finished" podID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerID="6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553" exitCode=0 Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.708336 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerDied","Data":"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553"} Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.708365 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"32d1d95a-1133-49b7-a59a-5a7687e4de2f","Type":"ContainerDied","Data":"98add4845fcda33a45dc700d5fd4ee34d1490812b9a1433f955c3aeca9b4198a"} Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.708379 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.708385 4759 scope.go:117] "RemoveContainer" containerID="6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.710987 4759 generic.go:334] "Generic (PLEG): container finished" podID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerID="d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a" exitCode=0 Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.711028 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerDied","Data":"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a"} Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.711071 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.711085 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555","Type":"ContainerDied","Data":"f9632e81ba3fbab710a1ab05cffdfc07cf5fbd3a783cdb7baf4fe5b54247b376"} Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.738223 4759 scope.go:117] "RemoveContainer" containerID="b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.741951 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.749282 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.755009 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.757208 4759 scope.go:117] "RemoveContainer" containerID="6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553" Nov 25 19:53:55 crc kubenswrapper[4759]: E1125 19:53:55.757737 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553\": container with ID starting with 6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553 not found: ID does not exist" containerID="6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.757771 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553"} err="failed to get container status \"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553\": rpc error: code = NotFound desc = could not find container \"6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553\": container with ID starting with 6956aff46cf2431cb71946f6e71ea74ff545fd63501a03cf91b3af7246bd6553 not found: ID does not exist" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.757797 4759 scope.go:117] "RemoveContainer" containerID="b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0" Nov 25 19:53:55 crc kubenswrapper[4759]: E1125 19:53:55.758102 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0\": container with ID starting with b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0 not found: ID does not exist" containerID="b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.758132 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0"} err="failed to get container status \"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0\": rpc error: code = NotFound desc = could not find container \"b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0\": container with ID starting with b6911a40f38304ccff844852aed4b01e7cf424644d221101e792a7b7079c93e0 not found: ID does not exist" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.758154 4759 scope.go:117] "RemoveContainer" containerID="d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.760988 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.778210 4759 scope.go:117] "RemoveContainer" containerID="8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.794659 4759 scope.go:117] "RemoveContainer" containerID="d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a" Nov 25 19:53:55 crc kubenswrapper[4759]: E1125 19:53:55.795092 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a\": container with ID starting with d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a not found: ID does not exist" containerID="d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.795117 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a"} err="failed to get container status \"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a\": rpc error: code = NotFound desc = could not find container \"d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a\": container with ID starting with d21f427815a62eb2ce494b4303ae25e4db3bdce3c7c62353625280a8cbf6074a not found: ID does not exist" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.795137 4759 scope.go:117] "RemoveContainer" containerID="8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f" Nov 25 19:53:55 crc kubenswrapper[4759]: E1125 19:53:55.795490 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f\": container with ID starting with 8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f not found: ID does not exist" containerID="8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f" Nov 25 19:53:55 crc kubenswrapper[4759]: I1125 19:53:55.795509 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f"} err="failed to get container status \"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f\": rpc error: code = NotFound desc = could not find container \"8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f\": container with ID starting with 8dd2d600208de27020e8ae32f7b38265e6db5e73f807a1f5d06b10fb23a4b78f not found: ID does not exist" Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.120112 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" path="/var/lib/kubelet/pods/32d1d95a-1133-49b7-a59a-5a7687e4de2f/volumes" Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.121492 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" path="/var/lib/kubelet/pods/aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555/volumes" Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.533026 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-6g2z8"] Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.538111 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-6g2z8"] Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.548540 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance5fe1-account-delete-bp2nk"] Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.554242 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2"] Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.559766 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-5fe1-account-create-update-vxgv2"] Nov 25 19:53:56 crc kubenswrapper[4759]: I1125 19:53:56.564108 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance5fe1-account-delete-bp2nk"] Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.119561 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46135ac8-9252-4497-8c18-480c1e0c9157" path="/var/lib/kubelet/pods/46135ac8-9252-4497-8c18-480c1e0c9157/volumes" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.120141 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="759d2f0f-0f92-4d86-b3a2-09301802f2a8" path="/var/lib/kubelet/pods/759d2f0f-0f92-4d86-b3a2-09301802f2a8/volumes" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.120693 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da3ff6ab-c18d-48cf-9f02-752df22a784e" path="/var/lib/kubelet/pods/da3ff6ab-c18d-48cf-9f02-752df22a784e/volumes" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139065 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-7jwjn"] Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139513 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" containerName="openstackclient" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139533 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" containerName="openstackclient" Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139545 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139554 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139570 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139576 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139585 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759d2f0f-0f92-4d86-b3a2-09301802f2a8" containerName="mariadb-account-delete" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139591 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="759d2f0f-0f92-4d86-b3a2-09301802f2a8" containerName="mariadb-account-delete" Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139602 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139608 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: E1125 19:53:58.139621 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139627 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139760 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139769 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="759d2f0f-0f92-4d86-b3a2-09301802f2a8" containerName="mariadb-account-delete" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139778 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139790 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbe3a07-6001-4b37-bb3f-bc31e30fcce6" containerName="openstackclient" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139802 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d1d95a-1133-49b7-a59a-5a7687e4de2f" containerName="glance-log" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.139811 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="aead7fbe-fa6f-4cd3-8c6f-e62e7a7e3555" containerName="glance-httpd" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.140266 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.145874 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk"] Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.146874 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.148523 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.151870 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-7jwjn"] Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.167264 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk"] Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.251386 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cxcv\" (UniqueName: \"kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.251570 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.251603 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvv2m\" (UniqueName: \"kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.251639 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.353587 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cxcv\" (UniqueName: \"kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.354075 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.354838 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvv2m\" (UniqueName: \"kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.354995 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.355013 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.355765 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.370684 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cxcv\" (UniqueName: \"kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv\") pod \"glance-db-create-7jwjn\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.371237 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvv2m\" (UniqueName: \"kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m\") pod \"glance-fd5b-account-create-update-8f4bk\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.462030 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.470494 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.727434 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk"] Nov 25 19:53:58 crc kubenswrapper[4759]: W1125 19:53:58.889546 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod536b557c_4798_4ba9_8d92_153e9c513d92.slice/crio-47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564 WatchSource:0}: Error finding container 47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564: Status 404 returned error can't find the container with id 47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564 Nov 25 19:53:58 crc kubenswrapper[4759]: I1125 19:53:58.889580 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-7jwjn"] Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.745718 4759 generic.go:334] "Generic (PLEG): container finished" podID="536b557c-4798-4ba9-8d92-153e9c513d92" containerID="6cab315d8ded72ac3a7d990022f75fefa10f1c422c74046180f779666a9e742f" exitCode=0 Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.745818 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-7jwjn" event={"ID":"536b557c-4798-4ba9-8d92-153e9c513d92","Type":"ContainerDied","Data":"6cab315d8ded72ac3a7d990022f75fefa10f1c422c74046180f779666a9e742f"} Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.745860 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-7jwjn" event={"ID":"536b557c-4798-4ba9-8d92-153e9c513d92","Type":"ContainerStarted","Data":"47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564"} Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.747196 4759 generic.go:334] "Generic (PLEG): container finished" podID="6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" containerID="c4554495742a389f8c8b9f0f25ae299ac86225f7b95f8111af1249defbda7acd" exitCode=0 Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.747253 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" event={"ID":"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f","Type":"ContainerDied","Data":"c4554495742a389f8c8b9f0f25ae299ac86225f7b95f8111af1249defbda7acd"} Nov 25 19:53:59 crc kubenswrapper[4759]: I1125 19:53:59.747284 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" event={"ID":"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f","Type":"ContainerStarted","Data":"3ab511f3083129accbd1a4f6fcff9e6bafdbac23e71c00ccae4e102f62f69a68"} Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.089722 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.093684 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.195356 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts\") pod \"536b557c-4798-4ba9-8d92-153e9c513d92\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.195408 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvv2m\" (UniqueName: \"kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m\") pod \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.195479 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cxcv\" (UniqueName: \"kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv\") pod \"536b557c-4798-4ba9-8d92-153e9c513d92\" (UID: \"536b557c-4798-4ba9-8d92-153e9c513d92\") " Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.195504 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts\") pod \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\" (UID: \"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f\") " Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.196043 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "536b557c-4798-4ba9-8d92-153e9c513d92" (UID: "536b557c-4798-4ba9-8d92-153e9c513d92"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.196055 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" (UID: "6c2c1ed8-ccd2-4828-95aa-18cc84e5315f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.200286 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv" (OuterVolumeSpecName: "kube-api-access-6cxcv") pod "536b557c-4798-4ba9-8d92-153e9c513d92" (UID: "536b557c-4798-4ba9-8d92-153e9c513d92"). InnerVolumeSpecName "kube-api-access-6cxcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.200562 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m" (OuterVolumeSpecName: "kube-api-access-rvv2m") pod "6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" (UID: "6c2c1ed8-ccd2-4828-95aa-18cc84e5315f"). InnerVolumeSpecName "kube-api-access-rvv2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.296885 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/536b557c-4798-4ba9-8d92-153e9c513d92-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.296912 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvv2m\" (UniqueName: \"kubernetes.io/projected/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-kube-api-access-rvv2m\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.296922 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cxcv\" (UniqueName: \"kubernetes.io/projected/536b557c-4798-4ba9-8d92-153e9c513d92-kube-api-access-6cxcv\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.296931 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.760430 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-7jwjn" event={"ID":"536b557c-4798-4ba9-8d92-153e9c513d92","Type":"ContainerDied","Data":"47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564"} Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.760484 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-7jwjn" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.760498 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47e06130611a972bb9a7d46593392515b644dc5123eeb48d9366c495885df564" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.762262 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" event={"ID":"6c2c1ed8-ccd2-4828-95aa-18cc84e5315f","Type":"ContainerDied","Data":"3ab511f3083129accbd1a4f6fcff9e6bafdbac23e71c00ccae4e102f62f69a68"} Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.762283 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ab511f3083129accbd1a4f6fcff9e6bafdbac23e71c00ccae4e102f62f69a68" Nov 25 19:54:01 crc kubenswrapper[4759]: I1125 19:54:01.762344 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.288882 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-xh8dg"] Nov 25 19:54:03 crc kubenswrapper[4759]: E1125 19:54:03.289436 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536b557c-4798-4ba9-8d92-153e9c513d92" containerName="mariadb-database-create" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.289467 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="536b557c-4798-4ba9-8d92-153e9c513d92" containerName="mariadb-database-create" Nov 25 19:54:03 crc kubenswrapper[4759]: E1125 19:54:03.289489 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" containerName="mariadb-account-create-update" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.289495 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" containerName="mariadb-account-create-update" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.289609 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="536b557c-4798-4ba9-8d92-153e9c513d92" containerName="mariadb-database-create" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.289621 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" containerName="mariadb-account-create-update" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.290027 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.293225 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-k6vh4" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.293827 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.293860 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.317827 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-xh8dg"] Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.361589 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.361652 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpxbt\" (UniqueName: \"kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.361754 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.361785 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.463056 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.463363 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpxbt\" (UniqueName: \"kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.463406 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.463428 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.468721 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.468771 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.468870 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.487517 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpxbt\" (UniqueName: \"kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt\") pod \"glance-db-sync-xh8dg\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:03 crc kubenswrapper[4759]: I1125 19:54:03.673011 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:04 crc kubenswrapper[4759]: I1125 19:54:04.104027 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-xh8dg"] Nov 25 19:54:04 crc kubenswrapper[4759]: I1125 19:54:04.785133 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-xh8dg" event={"ID":"48c33c22-0d76-4c74-b1f6-176d5cf2eec3","Type":"ContainerStarted","Data":"f5b621c8a24752948a9a7e746ea52014f05c59397538038fcc2e18aaa9d3193a"} Nov 25 19:54:04 crc kubenswrapper[4759]: I1125 19:54:04.785398 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-xh8dg" event={"ID":"48c33c22-0d76-4c74-b1f6-176d5cf2eec3","Type":"ContainerStarted","Data":"39f92e302a36757cdce5284635d457c41c8967fbf5468225376c88008d11f3a6"} Nov 25 19:54:04 crc kubenswrapper[4759]: I1125 19:54:04.804184 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-xh8dg" podStartSLOduration=1.804164077 podStartE2EDuration="1.804164077s" podCreationTimestamp="2025-11-25 19:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:04.800226678 +0000 UTC m=+1235.386174369" watchObservedRunningTime="2025-11-25 19:54:04.804164077 +0000 UTC m=+1235.390111788" Nov 25 19:54:07 crc kubenswrapper[4759]: I1125 19:54:07.809781 4759 generic.go:334] "Generic (PLEG): container finished" podID="48c33c22-0d76-4c74-b1f6-176d5cf2eec3" containerID="f5b621c8a24752948a9a7e746ea52014f05c59397538038fcc2e18aaa9d3193a" exitCode=0 Nov 25 19:54:07 crc kubenswrapper[4759]: I1125 19:54:07.809837 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-xh8dg" event={"ID":"48c33c22-0d76-4c74-b1f6-176d5cf2eec3","Type":"ContainerDied","Data":"f5b621c8a24752948a9a7e746ea52014f05c59397538038fcc2e18aaa9d3193a"} Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.078068 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.246680 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data\") pod \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.246753 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data\") pod \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.246781 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle\") pod \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.246813 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpxbt\" (UniqueName: \"kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt\") pod \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\" (UID: \"48c33c22-0d76-4c74-b1f6-176d5cf2eec3\") " Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.255215 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "48c33c22-0d76-4c74-b1f6-176d5cf2eec3" (UID: "48c33c22-0d76-4c74-b1f6-176d5cf2eec3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.255312 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt" (OuterVolumeSpecName: "kube-api-access-bpxbt") pod "48c33c22-0d76-4c74-b1f6-176d5cf2eec3" (UID: "48c33c22-0d76-4c74-b1f6-176d5cf2eec3"). InnerVolumeSpecName "kube-api-access-bpxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.266602 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48c33c22-0d76-4c74-b1f6-176d5cf2eec3" (UID: "48c33c22-0d76-4c74-b1f6-176d5cf2eec3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.290185 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data" (OuterVolumeSpecName: "config-data") pod "48c33c22-0d76-4c74-b1f6-176d5cf2eec3" (UID: "48c33c22-0d76-4c74-b1f6-176d5cf2eec3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.348397 4759 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.348431 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.348440 4759 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.348473 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpxbt\" (UniqueName: \"kubernetes.io/projected/48c33c22-0d76-4c74-b1f6-176d5cf2eec3-kube-api-access-bpxbt\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.828044 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-xh8dg" event={"ID":"48c33c22-0d76-4c74-b1f6-176d5cf2eec3","Type":"ContainerDied","Data":"39f92e302a36757cdce5284635d457c41c8967fbf5468225376c88008d11f3a6"} Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.828378 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39f92e302a36757cdce5284635d457c41c8967fbf5468225376c88008d11f3a6" Nov 25 19:54:09 crc kubenswrapper[4759]: I1125 19:54:09.828095 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-xh8dg" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.244209 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:10 crc kubenswrapper[4759]: E1125 19:54:10.244485 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c33c22-0d76-4c74-b1f6-176d5cf2eec3" containerName="glance-db-sync" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.244497 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c33c22-0d76-4c74-b1f6-176d5cf2eec3" containerName="glance-db-sync" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.244622 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c33c22-0d76-4c74-b1f6-176d5cf2eec3" containerName="glance-db-sync" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.245276 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.247098 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-k6vh4" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.247913 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.248299 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.248419 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.248508 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.252904 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.267152 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.362130 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.362624 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.362827 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.362909 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58tmc\" (UniqueName: \"kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.363100 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.363209 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.363253 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.363279 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.363555 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.464895 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465260 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465376 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465379 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465604 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465658 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465678 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465730 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465798 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58tmc\" (UniqueName: \"kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.465892 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.466199 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.466230 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.470119 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.470768 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.471503 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.471671 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.481779 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.485207 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58tmc\" (UniqueName: \"kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.486200 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.558406 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:10 crc kubenswrapper[4759]: I1125 19:54:10.836883 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.056467 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.862042 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerStarted","Data":"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19"} Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.862584 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerStarted","Data":"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430"} Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.862605 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerStarted","Data":"854affd1d825e704a42082e5984cdff18d60ce5c13c49fad08a5d79fa35ad98d"} Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.862333 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-log" containerID="cri-o://e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" gracePeriod=30 Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.862970 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-httpd" containerID="cri-o://9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" gracePeriod=30 Nov 25 19:54:11 crc kubenswrapper[4759]: I1125 19:54:11.893170 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=1.893149728 podStartE2EDuration="1.893149728s" podCreationTimestamp="2025-11-25 19:54:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:11.888561831 +0000 UTC m=+1242.474509552" watchObservedRunningTime="2025-11-25 19:54:11.893149728 +0000 UTC m=+1242.479097429" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.298041 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.390820 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.390906 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.390954 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.390997 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.391028 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58tmc\" (UniqueName: \"kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.391062 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.391082 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.391125 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.391157 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data\") pod \"14c0cb15-e959-4f52-b2e7-8da110936642\" (UID: \"14c0cb15-e959-4f52-b2e7-8da110936642\") " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.392481 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs" (OuterVolumeSpecName: "logs") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.392723 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.395717 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.395852 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc" (OuterVolumeSpecName: "kube-api-access-58tmc") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "kube-api-access-58tmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.396104 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts" (OuterVolumeSpecName: "scripts") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.410465 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.422538 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.424405 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data" (OuterVolumeSpecName: "config-data") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.424854 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14c0cb15-e959-4f52-b2e7-8da110936642" (UID: "14c0cb15-e959-4f52-b2e7-8da110936642"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492618 4759 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492648 4759 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492658 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58tmc\" (UniqueName: \"kubernetes.io/projected/14c0cb15-e959-4f52-b2e7-8da110936642-kube-api-access-58tmc\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492670 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492678 4759 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492686 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492694 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c0cb15-e959-4f52-b2e7-8da110936642-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492702 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c0cb15-e959-4f52-b2e7-8da110936642-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.492730 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.505620 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.593712 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871477 4759 generic.go:334] "Generic (PLEG): container finished" podID="14c0cb15-e959-4f52-b2e7-8da110936642" containerID="9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" exitCode=143 Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871522 4759 generic.go:334] "Generic (PLEG): container finished" podID="14c0cb15-e959-4f52-b2e7-8da110936642" containerID="e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" exitCode=143 Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871490 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871539 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerDied","Data":"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19"} Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871587 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerDied","Data":"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430"} Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871599 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"14c0cb15-e959-4f52-b2e7-8da110936642","Type":"ContainerDied","Data":"854affd1d825e704a42082e5984cdff18d60ce5c13c49fad08a5d79fa35ad98d"} Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.871615 4759 scope.go:117] "RemoveContainer" containerID="9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.891327 4759 scope.go:117] "RemoveContainer" containerID="e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.908386 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.916297 4759 scope.go:117] "RemoveContainer" containerID="9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.918284 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:12 crc kubenswrapper[4759]: E1125 19:54:12.920629 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19\": container with ID starting with 9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19 not found: ID does not exist" containerID="9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.920673 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19"} err="failed to get container status \"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19\": rpc error: code = NotFound desc = could not find container \"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19\": container with ID starting with 9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19 not found: ID does not exist" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.920700 4759 scope.go:117] "RemoveContainer" containerID="e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" Nov 25 19:54:12 crc kubenswrapper[4759]: E1125 19:54:12.921172 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430\": container with ID starting with e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430 not found: ID does not exist" containerID="e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.921192 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430"} err="failed to get container status \"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430\": rpc error: code = NotFound desc = could not find container \"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430\": container with ID starting with e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430 not found: ID does not exist" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.921207 4759 scope.go:117] "RemoveContainer" containerID="9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.921871 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19"} err="failed to get container status \"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19\": rpc error: code = NotFound desc = could not find container \"9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19\": container with ID starting with 9e17c90aaa63677fa7bc0697ae3c4d4ebff00d04486e6596243477afce749d19 not found: ID does not exist" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.921889 4759 scope.go:117] "RemoveContainer" containerID="e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.922141 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430"} err="failed to get container status \"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430\": rpc error: code = NotFound desc = could not find container \"e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430\": container with ID starting with e8670202422c0e3c8e4c27d2e218335bb8dfbec2ce64c1af98f3316987f56430 not found: ID does not exist" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.932034 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:12 crc kubenswrapper[4759]: E1125 19:54:12.932525 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-httpd" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.932549 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-httpd" Nov 25 19:54:12 crc kubenswrapper[4759]: E1125 19:54:12.932625 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-log" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.932635 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-log" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.933050 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-log" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.933082 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" containerName="glance-httpd" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.933927 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.937878 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.970135 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.970557 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.970700 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.970814 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.970841 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.971052 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-k6vh4" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999663 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fpmk\" (UniqueName: \"kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999732 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999783 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999837 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999874 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999904 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:12 crc kubenswrapper[4759]: I1125 19:54:12.999934 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.000069 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.000169 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.101757 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fpmk\" (UniqueName: \"kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102061 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102085 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102100 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102121 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102137 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102162 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102187 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102222 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102406 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102654 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.102781 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.107373 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.107994 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.109291 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.109422 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.114118 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.120183 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.124699 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fpmk\" (UniqueName: \"kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk\") pod \"glance-default-single-0\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.287641 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.713894 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:13 crc kubenswrapper[4759]: I1125 19:54:13.880825 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerStarted","Data":"d8e4928d73f08f10cd073a21c23ed7714913ddbd04e5f7f18b9a58461ecb166a"} Nov 25 19:54:14 crc kubenswrapper[4759]: I1125 19:54:14.120636 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c0cb15-e959-4f52-b2e7-8da110936642" path="/var/lib/kubelet/pods/14c0cb15-e959-4f52-b2e7-8da110936642/volumes" Nov 25 19:54:14 crc kubenswrapper[4759]: I1125 19:54:14.889258 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerStarted","Data":"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a"} Nov 25 19:54:14 crc kubenswrapper[4759]: I1125 19:54:14.889591 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerStarted","Data":"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071"} Nov 25 19:54:14 crc kubenswrapper[4759]: I1125 19:54:14.906527 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.906509629 podStartE2EDuration="2.906509629s" podCreationTimestamp="2025-11-25 19:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:14.905843311 +0000 UTC m=+1245.491791012" watchObservedRunningTime="2025-11-25 19:54:14.906509629 +0000 UTC m=+1245.492457340" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.288753 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.290102 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.318584 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.329918 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.958307 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:23 crc kubenswrapper[4759]: I1125 19:54:23.958349 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:24 crc kubenswrapper[4759]: I1125 19:54:24.896270 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:54:24 crc kubenswrapper[4759]: I1125 19:54:24.896662 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:54:24 crc kubenswrapper[4759]: I1125 19:54:24.896713 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:54:24 crc kubenswrapper[4759]: I1125 19:54:24.897334 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:54:24 crc kubenswrapper[4759]: I1125 19:54:24.897393 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930" gracePeriod=600 Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.868325 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.888761 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.981043 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930" exitCode=0 Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.981117 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930"} Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.982050 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2"} Nov 25 19:54:25 crc kubenswrapper[4759]: I1125 19:54:25.982089 4759 scope.go:117] "RemoveContainer" containerID="eebf0f70d5823717fb0d81520aa7583efa3d223c13e5a101fbbb2ce42fb317e8" Nov 25 19:54:26 crc kubenswrapper[4759]: I1125 19:54:26.924990 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-xh8dg"] Nov 25 19:54:26 crc kubenswrapper[4759]: I1125 19:54:26.930800 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-xh8dg"] Nov 25 19:54:26 crc kubenswrapper[4759]: I1125 19:54:26.992322 4759 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/glance-default-single-0" secret="" err="secret \"glance-glance-dockercfg-k6vh4\" not found" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.003186 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancefd5b-account-delete-kc7zh"] Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.004260 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.026720 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancefd5b-account-delete-kc7zh"] Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.055557 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.107797 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.107870 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr5gs\" (UniqueName: \"kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.108093 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.108150 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:27.608132336 +0000 UTC m=+1258.194080037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-default-single-config-data" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.108411 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.108504 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:27.608482835 +0000 UTC m=+1258.194430536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-scripts" not found Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.209676 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.209730 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr5gs\" (UniqueName: \"kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.210698 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.233492 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr5gs\" (UniqueName: \"kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs\") pod \"glancefd5b-account-delete-kc7zh\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.318907 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.615913 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.615929 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.616004 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:28.615989086 +0000 UTC m=+1259.201936787 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-scripts" not found Nov 25 19:54:27 crc kubenswrapper[4759]: E1125 19:54:27.616094 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:28.616017987 +0000 UTC m=+1259.201965728 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-default-single-config-data" not found Nov 25 19:54:27 crc kubenswrapper[4759]: I1125 19:54:27.719618 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancefd5b-account-delete-kc7zh"] Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.001104 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" event={"ID":"dacc22e7-1175-4d8a-8aa9-4403f015b588","Type":"ContainerStarted","Data":"9e7ab002ae6e5148facfe93c5b35ebacb9331334114aa2e9059f7a1abbe5ae1f"} Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.001538 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-log" containerID="cri-o://f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071" gracePeriod=30 Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.001619 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" event={"ID":"dacc22e7-1175-4d8a-8aa9-4403f015b588","Type":"ContainerStarted","Data":"6b09c7f3a4966d2aba05677ae70881f6cea0ff7b43325e3aa75f7a79e4ae3526"} Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.001561 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-httpd" containerID="cri-o://7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a" gracePeriod=30 Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.022225 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" podStartSLOduration=2.022205394 podStartE2EDuration="2.022205394s" podCreationTimestamp="2025-11-25 19:54:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:28.021842193 +0000 UTC m=+1258.607789934" watchObservedRunningTime="2025-11-25 19:54:28.022205394 +0000 UTC m=+1258.608153095" Nov 25 19:54:28 crc kubenswrapper[4759]: I1125 19:54:28.116676 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48c33c22-0d76-4c74-b1f6-176d5cf2eec3" path="/var/lib/kubelet/pods/48c33c22-0d76-4c74-b1f6-176d5cf2eec3/volumes" Nov 25 19:54:28 crc kubenswrapper[4759]: E1125 19:54:28.632712 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 19:54:28 crc kubenswrapper[4759]: E1125 19:54:28.632779 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:30.632762555 +0000 UTC m=+1261.218710256 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-default-single-config-data" not found Nov 25 19:54:28 crc kubenswrapper[4759]: E1125 19:54:28.632854 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 19:54:28 crc kubenswrapper[4759]: E1125 19:54:28.632926 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:30.632908249 +0000 UTC m=+1261.218855950 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-scripts" not found Nov 25 19:54:29 crc kubenswrapper[4759]: I1125 19:54:29.012569 4759 generic.go:334] "Generic (PLEG): container finished" podID="dacc22e7-1175-4d8a-8aa9-4403f015b588" containerID="9e7ab002ae6e5148facfe93c5b35ebacb9331334114aa2e9059f7a1abbe5ae1f" exitCode=0 Nov 25 19:54:29 crc kubenswrapper[4759]: I1125 19:54:29.012650 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" event={"ID":"dacc22e7-1175-4d8a-8aa9-4403f015b588","Type":"ContainerDied","Data":"9e7ab002ae6e5148facfe93c5b35ebacb9331334114aa2e9059f7a1abbe5ae1f"} Nov 25 19:54:29 crc kubenswrapper[4759]: I1125 19:54:29.015818 4759 generic.go:334] "Generic (PLEG): container finished" podID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerID="f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071" exitCode=143 Nov 25 19:54:29 crc kubenswrapper[4759]: I1125 19:54:29.015932 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerDied","Data":"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071"} Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.383475 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.466136 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr5gs\" (UniqueName: \"kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs\") pod \"dacc22e7-1175-4d8a-8aa9-4403f015b588\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.466218 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts\") pod \"dacc22e7-1175-4d8a-8aa9-4403f015b588\" (UID: \"dacc22e7-1175-4d8a-8aa9-4403f015b588\") " Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.466867 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dacc22e7-1175-4d8a-8aa9-4403f015b588" (UID: "dacc22e7-1175-4d8a-8aa9-4403f015b588"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.471212 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs" (OuterVolumeSpecName: "kube-api-access-nr5gs") pod "dacc22e7-1175-4d8a-8aa9-4403f015b588" (UID: "dacc22e7-1175-4d8a-8aa9-4403f015b588"). InnerVolumeSpecName "kube-api-access-nr5gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.567863 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr5gs\" (UniqueName: \"kubernetes.io/projected/dacc22e7-1175-4d8a-8aa9-4403f015b588-kube-api-access-nr5gs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:30 crc kubenswrapper[4759]: I1125 19:54:30.567887 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dacc22e7-1175-4d8a-8aa9-4403f015b588-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:30 crc kubenswrapper[4759]: E1125 19:54:30.669175 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 19:54:30 crc kubenswrapper[4759]: E1125 19:54:30.669193 4759 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 25 19:54:30 crc kubenswrapper[4759]: E1125 19:54:30.669240 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:34.669224813 +0000 UTC m=+1265.255172514 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-scripts" not found Nov 25 19:54:30 crc kubenswrapper[4759]: E1125 19:54:30.669284 4759 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data podName:a613b966-bc42-4649-82eb-f7d807f18fbc nodeName:}" failed. No retries permitted until 2025-11-25 19:54:34.669263824 +0000 UTC m=+1265.255211535 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data") pod "glance-default-single-0" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc") : secret "glance-default-single-config-data" not found Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.041525 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" event={"ID":"dacc22e7-1175-4d8a-8aa9-4403f015b588","Type":"ContainerDied","Data":"6b09c7f3a4966d2aba05677ae70881f6cea0ff7b43325e3aa75f7a79e4ae3526"} Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.041565 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b09c7f3a4966d2aba05677ae70881f6cea0ff7b43325e3aa75f7a79e4ae3526" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.041662 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefd5b-account-delete-kc7zh" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.500559 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.596846 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.596882 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.596971 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.596992 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597017 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fpmk\" (UniqueName: \"kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597056 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597099 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597120 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597140 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data\") pod \"a613b966-bc42-4649-82eb-f7d807f18fbc\" (UID: \"a613b966-bc42-4649-82eb-f7d807f18fbc\") " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.597775 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.598015 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs" (OuterVolumeSpecName: "logs") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.601241 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk" (OuterVolumeSpecName: "kube-api-access-2fpmk") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "kube-api-access-2fpmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.601618 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts" (OuterVolumeSpecName: "scripts") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.615311 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.617300 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.630671 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data" (OuterVolumeSpecName: "config-data") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.633600 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.641683 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a613b966-bc42-4649-82eb-f7d807f18fbc" (UID: "a613b966-bc42-4649-82eb-f7d807f18fbc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698118 4759 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698230 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698298 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fpmk\" (UniqueName: \"kubernetes.io/projected/a613b966-bc42-4649-82eb-f7d807f18fbc-kube-api-access-2fpmk\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698371 4759 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698421 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a613b966-bc42-4649-82eb-f7d807f18fbc-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698527 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698597 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698677 4759 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.698729 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a613b966-bc42-4649-82eb-f7d807f18fbc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.716636 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 19:54:31 crc kubenswrapper[4759]: I1125 19:54:31.800640 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.024560 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-7jwjn"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.036892 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-7jwjn"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.044643 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancefd5b-account-delete-kc7zh"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.050332 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.052039 4759 generic.go:334] "Generic (PLEG): container finished" podID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerID="7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a" exitCode=0 Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.052078 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerDied","Data":"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a"} Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.052108 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a613b966-bc42-4649-82eb-f7d807f18fbc","Type":"ContainerDied","Data":"d8e4928d73f08f10cd073a21c23ed7714913ddbd04e5f7f18b9a58461ecb166a"} Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.052124 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.052153 4759 scope.go:117] "RemoveContainer" containerID="7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.057132 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-fd5b-account-create-update-8f4bk"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.064534 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancefd5b-account-delete-kc7zh"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.083062 4759 scope.go:117] "RemoveContainer" containerID="f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.095740 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.104117 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.107713 4759 scope.go:117] "RemoveContainer" containerID="7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a" Nov 25 19:54:32 crc kubenswrapper[4759]: E1125 19:54:32.108241 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a\": container with ID starting with 7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a not found: ID does not exist" containerID="7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.108285 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a"} err="failed to get container status \"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a\": rpc error: code = NotFound desc = could not find container \"7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a\": container with ID starting with 7b317352308245d1c18f49a509c86ae40e6171ee1d8e25ddebf1fbb2d434bd4a not found: ID does not exist" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.108312 4759 scope.go:117] "RemoveContainer" containerID="f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071" Nov 25 19:54:32 crc kubenswrapper[4759]: E1125 19:54:32.109041 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071\": container with ID starting with f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071 not found: ID does not exist" containerID="f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.109067 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071"} err="failed to get container status \"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071\": rpc error: code = NotFound desc = could not find container \"f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071\": container with ID starting with f59be9e3d2583571749413d653dfe0dde80b6eb5c4b0dd21cd6871a1b0989071 not found: ID does not exist" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.121790 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536b557c-4798-4ba9-8d92-153e9c513d92" path="/var/lib/kubelet/pods/536b557c-4798-4ba9-8d92-153e9c513d92/volumes" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.122854 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c2c1ed8-ccd2-4828-95aa-18cc84e5315f" path="/var/lib/kubelet/pods/6c2c1ed8-ccd2-4828-95aa-18cc84e5315f/volumes" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.123673 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" path="/var/lib/kubelet/pods/a613b966-bc42-4649-82eb-f7d807f18fbc/volumes" Nov 25 19:54:32 crc kubenswrapper[4759]: I1125 19:54:32.125046 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dacc22e7-1175-4d8a-8aa9-4403f015b588" path="/var/lib/kubelet/pods/dacc22e7-1175-4d8a-8aa9-4403f015b588/volumes" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.037942 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-lvtvr"] Nov 25 19:54:33 crc kubenswrapper[4759]: E1125 19:54:33.038326 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-httpd" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038346 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-httpd" Nov 25 19:54:33 crc kubenswrapper[4759]: E1125 19:54:33.038372 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dacc22e7-1175-4d8a-8aa9-4403f015b588" containerName="mariadb-account-delete" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038386 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacc22e7-1175-4d8a-8aa9-4403f015b588" containerName="mariadb-account-delete" Nov 25 19:54:33 crc kubenswrapper[4759]: E1125 19:54:33.038415 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-log" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038428 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-log" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038652 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-httpd" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038680 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="dacc22e7-1175-4d8a-8aa9-4403f015b588" containerName="mariadb-account-delete" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.038707 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="a613b966-bc42-4649-82eb-f7d807f18fbc" containerName="glance-log" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.039354 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.049603 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-lvtvr"] Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.056875 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-6696-account-create-update-pxz2s"] Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.057859 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.060736 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.061790 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6696-account-create-update-pxz2s"] Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.123142 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.123346 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxd7l\" (UniqueName: \"kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.225211 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.225283 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.225970 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.226196 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc5gt\" (UniqueName: \"kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.226689 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxd7l\" (UniqueName: \"kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.253541 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxd7l\" (UniqueName: \"kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l\") pod \"glance-db-create-lvtvr\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.328127 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.328247 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc5gt\" (UniqueName: \"kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.329232 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.345023 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc5gt\" (UniqueName: \"kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt\") pod \"glance-6696-account-create-update-pxz2s\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.367691 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.382563 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.676094 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-lvtvr"] Nov 25 19:54:33 crc kubenswrapper[4759]: W1125 19:54:33.682655 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f410cd7_ce9c_4df1_a0b6_32b54fe82bdf.slice/crio-9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800 WatchSource:0}: Error finding container 9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800: Status 404 returned error can't find the container with id 9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800 Nov 25 19:54:33 crc kubenswrapper[4759]: I1125 19:54:33.807319 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6696-account-create-update-pxz2s"] Nov 25 19:54:34 crc kubenswrapper[4759]: I1125 19:54:34.088392 4759 generic.go:334] "Generic (PLEG): container finished" podID="7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" containerID="909590ff9c2c5257bc0400f06753c5963581c0cecf55110a06839ee6dd296abf" exitCode=0 Nov 25 19:54:34 crc kubenswrapper[4759]: I1125 19:54:34.088502 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-lvtvr" event={"ID":"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf","Type":"ContainerDied","Data":"909590ff9c2c5257bc0400f06753c5963581c0cecf55110a06839ee6dd296abf"} Nov 25 19:54:34 crc kubenswrapper[4759]: I1125 19:54:34.088829 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-lvtvr" event={"ID":"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf","Type":"ContainerStarted","Data":"9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800"} Nov 25 19:54:34 crc kubenswrapper[4759]: I1125 19:54:34.092556 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" event={"ID":"965de3f2-28e4-4190-8528-305cd8978d33","Type":"ContainerStarted","Data":"372db637c6707ae6010d0dca2eca2783cd94a1989227a8f88c8a1ab20f8140b5"} Nov 25 19:54:34 crc kubenswrapper[4759]: I1125 19:54:34.092699 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" event={"ID":"965de3f2-28e4-4190-8528-305cd8978d33","Type":"ContainerStarted","Data":"9220c6275ec1e386342933c3fe697f08063052f215fd7b8bca73a8e00d350566"} Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.106110 4759 generic.go:334] "Generic (PLEG): container finished" podID="965de3f2-28e4-4190-8528-305cd8978d33" containerID="372db637c6707ae6010d0dca2eca2783cd94a1989227a8f88c8a1ab20f8140b5" exitCode=0 Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.106171 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" event={"ID":"965de3f2-28e4-4190-8528-305cd8978d33","Type":"ContainerDied","Data":"372db637c6707ae6010d0dca2eca2783cd94a1989227a8f88c8a1ab20f8140b5"} Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.555655 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.692785 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts\") pod \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.693248 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxd7l\" (UniqueName: \"kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l\") pod \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\" (UID: \"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf\") " Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.693467 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" (UID: "7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.693720 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.698826 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l" (OuterVolumeSpecName: "kube-api-access-jxd7l") pod "7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" (UID: "7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf"). InnerVolumeSpecName "kube-api-access-jxd7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:35 crc kubenswrapper[4759]: I1125 19:54:35.794870 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxd7l\" (UniqueName: \"kubernetes.io/projected/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf-kube-api-access-jxd7l\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.115684 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-lvtvr" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.121178 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-lvtvr" event={"ID":"7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf","Type":"ContainerDied","Data":"9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800"} Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.121249 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9227a8f2ea2c4a1346cf6563729ed4036259be59334201608f3a35ff6c231800" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.511313 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.613489 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc5gt\" (UniqueName: \"kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt\") pod \"965de3f2-28e4-4190-8528-305cd8978d33\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.613755 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts\") pod \"965de3f2-28e4-4190-8528-305cd8978d33\" (UID: \"965de3f2-28e4-4190-8528-305cd8978d33\") " Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.614624 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "965de3f2-28e4-4190-8528-305cd8978d33" (UID: "965de3f2-28e4-4190-8528-305cd8978d33"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.617135 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt" (OuterVolumeSpecName: "kube-api-access-kc5gt") pod "965de3f2-28e4-4190-8528-305cd8978d33" (UID: "965de3f2-28e4-4190-8528-305cd8978d33"). InnerVolumeSpecName "kube-api-access-kc5gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.716369 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/965de3f2-28e4-4190-8528-305cd8978d33-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:36 crc kubenswrapper[4759]: I1125 19:54:36.716407 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc5gt\" (UniqueName: \"kubernetes.io/projected/965de3f2-28e4-4190-8528-305cd8978d33-kube-api-access-kc5gt\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:37 crc kubenswrapper[4759]: I1125 19:54:37.123565 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" Nov 25 19:54:37 crc kubenswrapper[4759]: I1125 19:54:37.123543 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6696-account-create-update-pxz2s" event={"ID":"965de3f2-28e4-4190-8528-305cd8978d33","Type":"ContainerDied","Data":"9220c6275ec1e386342933c3fe697f08063052f215fd7b8bca73a8e00d350566"} Nov 25 19:54:37 crc kubenswrapper[4759]: I1125 19:54:37.124046 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9220c6275ec1e386342933c3fe697f08063052f215fd7b8bca73a8e00d350566" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.202081 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-wxlrw"] Nov 25 19:54:38 crc kubenswrapper[4759]: E1125 19:54:38.204131 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" containerName="mariadb-database-create" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.204361 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" containerName="mariadb-database-create" Nov 25 19:54:38 crc kubenswrapper[4759]: E1125 19:54:38.204699 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="965de3f2-28e4-4190-8528-305cd8978d33" containerName="mariadb-account-create-update" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.204839 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="965de3f2-28e4-4190-8528-305cd8978d33" containerName="mariadb-account-create-update" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.205183 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="965de3f2-28e4-4190-8528-305cd8978d33" containerName="mariadb-account-create-update" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.205300 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" containerName="mariadb-database-create" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.206173 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.215082 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wjxfz" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.215545 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.216411 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wxlrw"] Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.344102 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.344165 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsw52\" (UniqueName: \"kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.344217 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.445830 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.446822 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsw52\" (UniqueName: \"kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.446923 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.455856 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.456057 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.478627 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsw52\" (UniqueName: \"kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52\") pod \"glance-db-sync-wxlrw\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:38 crc kubenswrapper[4759]: I1125 19:54:38.549252 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:39 crc kubenswrapper[4759]: I1125 19:54:39.005135 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wxlrw"] Nov 25 19:54:39 crc kubenswrapper[4759]: I1125 19:54:39.139410 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wxlrw" event={"ID":"8e78b547-f232-41e3-a1b3-df5415fec073","Type":"ContainerStarted","Data":"148a2fec327547d9f461e159095333711951109639a8d631f8a9008a3ebc2b18"} Nov 25 19:54:40 crc kubenswrapper[4759]: I1125 19:54:40.150986 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wxlrw" event={"ID":"8e78b547-f232-41e3-a1b3-df5415fec073","Type":"ContainerStarted","Data":"837ffc79cf7e9d10c681676a19def92de843ae76b9d608076740c1bc3beab99b"} Nov 25 19:54:40 crc kubenswrapper[4759]: I1125 19:54:40.180131 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-wxlrw" podStartSLOduration=2.18010674 podStartE2EDuration="2.18010674s" podCreationTimestamp="2025-11-25 19:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:40.16776065 +0000 UTC m=+1270.753708351" watchObservedRunningTime="2025-11-25 19:54:40.18010674 +0000 UTC m=+1270.766054471" Nov 25 19:54:42 crc kubenswrapper[4759]: I1125 19:54:42.168887 4759 generic.go:334] "Generic (PLEG): container finished" podID="8e78b547-f232-41e3-a1b3-df5415fec073" containerID="837ffc79cf7e9d10c681676a19def92de843ae76b9d608076740c1bc3beab99b" exitCode=0 Nov 25 19:54:42 crc kubenswrapper[4759]: I1125 19:54:42.168967 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wxlrw" event={"ID":"8e78b547-f232-41e3-a1b3-df5415fec073","Type":"ContainerDied","Data":"837ffc79cf7e9d10c681676a19def92de843ae76b9d608076740c1bc3beab99b"} Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.494689 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.617872 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data\") pod \"8e78b547-f232-41e3-a1b3-df5415fec073\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.617981 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsw52\" (UniqueName: \"kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52\") pod \"8e78b547-f232-41e3-a1b3-df5415fec073\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.618004 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data\") pod \"8e78b547-f232-41e3-a1b3-df5415fec073\" (UID: \"8e78b547-f232-41e3-a1b3-df5415fec073\") " Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.624303 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8e78b547-f232-41e3-a1b3-df5415fec073" (UID: "8e78b547-f232-41e3-a1b3-df5415fec073"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.624619 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52" (OuterVolumeSpecName: "kube-api-access-bsw52") pod "8e78b547-f232-41e3-a1b3-df5415fec073" (UID: "8e78b547-f232-41e3-a1b3-df5415fec073"). InnerVolumeSpecName "kube-api-access-bsw52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.653316 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data" (OuterVolumeSpecName: "config-data") pod "8e78b547-f232-41e3-a1b3-df5415fec073" (UID: "8e78b547-f232-41e3-a1b3-df5415fec073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.719615 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsw52\" (UniqueName: \"kubernetes.io/projected/8e78b547-f232-41e3-a1b3-df5415fec073-kube-api-access-bsw52\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.719655 4759 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:43 crc kubenswrapper[4759]: I1125 19:54:43.719664 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e78b547-f232-41e3-a1b3-df5415fec073-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:44 crc kubenswrapper[4759]: I1125 19:54:44.189858 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wxlrw" event={"ID":"8e78b547-f232-41e3-a1b3-df5415fec073","Type":"ContainerDied","Data":"148a2fec327547d9f461e159095333711951109639a8d631f8a9008a3ebc2b18"} Nov 25 19:54:44 crc kubenswrapper[4759]: I1125 19:54:44.189898 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="148a2fec327547d9f461e159095333711951109639a8d631f8a9008a3ebc2b18" Nov 25 19:54:44 crc kubenswrapper[4759]: I1125 19:54:44.190047 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wxlrw" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.479674 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 19:54:45 crc kubenswrapper[4759]: E1125 19:54:45.481316 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e78b547-f232-41e3-a1b3-df5415fec073" containerName="glance-db-sync" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.481357 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e78b547-f232-41e3-a1b3-df5415fec073" containerName="glance-db-sync" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.482464 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e78b547-f232-41e3-a1b3-df5415fec073" containerName="glance-db-sync" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.486372 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.489058 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.493249 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wjxfz" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.504029 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.548664 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.579466 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.580844 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.585283 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.607101 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650240 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650303 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650326 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650347 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650382 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650400 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nvtv\" (UniqueName: \"kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650433 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650544 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650564 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650615 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650662 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650724 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650761 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.650798 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752541 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752603 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752633 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752653 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752676 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752682 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752690 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752717 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752738 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752757 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752793 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752844 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752883 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752906 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752925 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nvtv\" (UniqueName: \"kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752951 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752966 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752984 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.752986 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753005 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7gt9\" (UniqueName: \"kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753024 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753045 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753061 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753078 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753093 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753110 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753126 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753147 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753167 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753182 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753205 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753225 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753248 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753330 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753426 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753582 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753811 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.753839 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.754796 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.760062 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.762815 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.776474 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.779524 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nvtv\" (UniqueName: \"kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.788536 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854347 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854396 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854432 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854479 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854498 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854522 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854547 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854573 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854591 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854641 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7gt9\" (UniqueName: \"kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854659 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854660 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854688 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854709 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854731 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854869 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854973 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854998 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.854651 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855345 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855293 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855311 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855264 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855424 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855514 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.855773 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.860761 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.862023 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.874306 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7gt9\" (UniqueName: \"kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.875627 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.899510 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:45 crc kubenswrapper[4759]: I1125 19:54:45.910014 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:46 crc kubenswrapper[4759]: I1125 19:54:46.257868 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:46 crc kubenswrapper[4759]: I1125 19:54:46.309077 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 19:54:46 crc kubenswrapper[4759]: I1125 19:54:46.375159 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:46 crc kubenswrapper[4759]: W1125 19:54:46.376520 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25b5bd0c_63b3_462f_b7ac_84767bdb3391.slice/crio-0df7a57d3455bc3d9df4588e6ce5215cd0209a83d5873a0562569d29c473a277 WatchSource:0}: Error finding container 0df7a57d3455bc3d9df4588e6ce5215cd0209a83d5873a0562569d29c473a277: Status 404 returned error can't find the container with id 0df7a57d3455bc3d9df4588e6ce5215cd0209a83d5873a0562569d29c473a277 Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.223202 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerStarted","Data":"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.223721 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerStarted","Data":"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.223735 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerStarted","Data":"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.223744 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerStarted","Data":"cc2e389b12415c48e3c84f6618716442a89d3b83a6b6b9608550a18e0c947c1e"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225175 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerStarted","Data":"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225220 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerStarted","Data":"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225237 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerStarted","Data":"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225251 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerStarted","Data":"0df7a57d3455bc3d9df4588e6ce5215cd0209a83d5873a0562569d29c473a277"} Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225393 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-log" containerID="cri-o://ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" gracePeriod=30 Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225403 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-api" containerID="cri-o://c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" gracePeriod=30 Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.225392 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-httpd" containerID="cri-o://bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" gracePeriod=30 Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.284234 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.284205707 podStartE2EDuration="2.284205707s" podCreationTimestamp="2025-11-25 19:54:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:47.27085658 +0000 UTC m=+1277.856804291" watchObservedRunningTime="2025-11-25 19:54:47.284205707 +0000 UTC m=+1277.870153408" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.301170 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.301155133 podStartE2EDuration="3.301155133s" podCreationTimestamp="2025-11-25 19:54:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:47.300844283 +0000 UTC m=+1277.886791984" watchObservedRunningTime="2025-11-25 19:54:47.301155133 +0000 UTC m=+1277.887102824" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.609743 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682301 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682646 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682670 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682709 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682746 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682778 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682806 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7gt9\" (UniqueName: \"kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682832 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682848 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682863 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682884 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682951 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682975 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.682993 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme\") pod \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\" (UID: \"25b5bd0c-63b3-462f-b7ac-84767bdb3391\") " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.683301 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.683340 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev" (OuterVolumeSpecName: "dev") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.683364 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run" (OuterVolumeSpecName: "run") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684161 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys" (OuterVolumeSpecName: "sys") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684416 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs" (OuterVolumeSpecName: "logs") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684436 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684479 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684495 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.684513 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.687550 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.688131 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts" (OuterVolumeSpecName: "scripts") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.688468 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.688492 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9" (OuterVolumeSpecName: "kube-api-access-k7gt9") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "kube-api-access-k7gt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.745054 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data" (OuterVolumeSpecName: "config-data") pod "25b5bd0c-63b3-462f-b7ac-84767bdb3391" (UID: "25b5bd0c-63b3-462f-b7ac-84767bdb3391"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785258 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785297 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785306 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785315 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785327 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785358 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785371 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785379 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785388 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25b5bd0c-63b3-462f-b7ac-84767bdb3391-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785397 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7gt9\" (UniqueName: \"kubernetes.io/projected/25b5bd0c-63b3-462f-b7ac-84767bdb3391-kube-api-access-k7gt9\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785408 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785416 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785425 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25b5bd0c-63b3-462f-b7ac-84767bdb3391-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.785434 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/25b5bd0c-63b3-462f-b7ac-84767bdb3391-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.798415 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.811603 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.886616 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:47 crc kubenswrapper[4759]: I1125 19:54:47.886659 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235152 4759 generic.go:334] "Generic (PLEG): container finished" podID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" exitCode=143 Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235529 4759 generic.go:334] "Generic (PLEG): container finished" podID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" exitCode=143 Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235552 4759 generic.go:334] "Generic (PLEG): container finished" podID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" exitCode=143 Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235204 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235221 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerDied","Data":"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760"} Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235670 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerDied","Data":"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54"} Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235697 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerDied","Data":"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79"} Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235707 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"25b5bd0c-63b3-462f-b7ac-84767bdb3391","Type":"ContainerDied","Data":"0df7a57d3455bc3d9df4588e6ce5215cd0209a83d5873a0562569d29c473a277"} Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.235724 4759 scope.go:117] "RemoveContainer" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.262743 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.265499 4759 scope.go:117] "RemoveContainer" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.270511 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282531 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.282806 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-httpd" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282827 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-httpd" Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.282839 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-log" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282848 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-log" Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.282858 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-api" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282865 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-api" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282983 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-log" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282992 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-httpd" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.282999 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" containerName="glance-api" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.284809 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.287868 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.295117 4759 scope.go:117] "RemoveContainer" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.305840 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.328318 4759 scope.go:117] "RemoveContainer" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.328890 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": container with ID starting with c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760 not found: ID does not exist" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.328941 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760"} err="failed to get container status \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": rpc error: code = NotFound desc = could not find container \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": container with ID starting with c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.328979 4759 scope.go:117] "RemoveContainer" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.329387 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": container with ID starting with bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54 not found: ID does not exist" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329420 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54"} err="failed to get container status \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": rpc error: code = NotFound desc = could not find container \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": container with ID starting with bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329440 4759 scope.go:117] "RemoveContainer" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" Nov 25 19:54:48 crc kubenswrapper[4759]: E1125 19:54:48.329678 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": container with ID starting with ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79 not found: ID does not exist" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329713 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79"} err="failed to get container status \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": rpc error: code = NotFound desc = could not find container \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": container with ID starting with ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329732 4759 scope.go:117] "RemoveContainer" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329925 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760"} err="failed to get container status \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": rpc error: code = NotFound desc = could not find container \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": container with ID starting with c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.329964 4759 scope.go:117] "RemoveContainer" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.330330 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54"} err="failed to get container status \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": rpc error: code = NotFound desc = could not find container \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": container with ID starting with bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.330359 4759 scope.go:117] "RemoveContainer" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.330698 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79"} err="failed to get container status \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": rpc error: code = NotFound desc = could not find container \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": container with ID starting with ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.330723 4759 scope.go:117] "RemoveContainer" containerID="c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.331259 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760"} err="failed to get container status \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": rpc error: code = NotFound desc = could not find container \"c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760\": container with ID starting with c9458ada92244e1a8f71257c4c69239eabc48da7ae42ba49e669d07e22be0760 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.331294 4759 scope.go:117] "RemoveContainer" containerID="bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.331948 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54"} err="failed to get container status \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": rpc error: code = NotFound desc = could not find container \"bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54\": container with ID starting with bd18042d3a6cf1ef1af90e744b7529c4c5968e3400b693a7760750a042a99c54 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.331979 4759 scope.go:117] "RemoveContainer" containerID="ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.332234 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79"} err="failed to get container status \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": rpc error: code = NotFound desc = could not find container \"ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79\": container with ID starting with ee4cc755124ffe9bf5f2e35e29799089f4d3c1c904930550c28465e1e33b5d79 not found: ID does not exist" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.392952 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbwhd\" (UniqueName: \"kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.392993 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393010 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393024 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393040 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393076 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393099 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393156 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393184 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393211 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393242 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393278 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393316 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.393350 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.494883 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.494946 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.494985 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495020 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495023 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495055 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495146 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495198 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495297 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbwhd\" (UniqueName: \"kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495318 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495347 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495368 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495400 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495479 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495510 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.495927 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496062 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496096 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496122 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496300 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496358 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496431 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.496502 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.498337 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.498566 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.505584 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.510834 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.513301 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbwhd\" (UniqueName: \"kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.519412 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.536552 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:48 crc kubenswrapper[4759]: I1125 19:54:48.614388 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:49 crc kubenswrapper[4759]: I1125 19:54:49.014337 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 19:54:49 crc kubenswrapper[4759]: W1125 19:54:49.027981 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11fcc57f_260e_4cdd_a2a5_acfce9c9b676.slice/crio-33e47a7835419fd7c0f3128c861d7ddd973c352de46ad21931b88084a597183e WatchSource:0}: Error finding container 33e47a7835419fd7c0f3128c861d7ddd973c352de46ad21931b88084a597183e: Status 404 returned error can't find the container with id 33e47a7835419fd7c0f3128c861d7ddd973c352de46ad21931b88084a597183e Nov 25 19:54:49 crc kubenswrapper[4759]: I1125 19:54:49.250351 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerStarted","Data":"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591"} Nov 25 19:54:49 crc kubenswrapper[4759]: I1125 19:54:49.250819 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerStarted","Data":"33e47a7835419fd7c0f3128c861d7ddd973c352de46ad21931b88084a597183e"} Nov 25 19:54:50 crc kubenswrapper[4759]: I1125 19:54:50.117351 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b5bd0c-63b3-462f-b7ac-84767bdb3391" path="/var/lib/kubelet/pods/25b5bd0c-63b3-462f-b7ac-84767bdb3391/volumes" Nov 25 19:54:50 crc kubenswrapper[4759]: I1125 19:54:50.259518 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerStarted","Data":"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947"} Nov 25 19:54:50 crc kubenswrapper[4759]: I1125 19:54:50.259574 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerStarted","Data":"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96"} Nov 25 19:54:50 crc kubenswrapper[4759]: I1125 19:54:50.296871 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.296603982 podStartE2EDuration="2.296603982s" podCreationTimestamp="2025-11-25 19:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:54:50.281992751 +0000 UTC m=+1280.867940462" watchObservedRunningTime="2025-11-25 19:54:50.296603982 +0000 UTC m=+1280.882551713" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.856684 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.858219 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.858301 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.879410 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.881307 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:55 crc kubenswrapper[4759]: I1125 19:54:55.891068 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.319055 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.319388 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.319399 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.332501 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.335784 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:56 crc kubenswrapper[4759]: I1125 19:54:56.345058 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.615426 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.615838 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.615862 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.663025 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.666104 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:58 crc kubenswrapper[4759]: I1125 19:54:58.675824 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.358503 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.358571 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.358583 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.370160 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.373397 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:54:59 crc kubenswrapper[4759]: I1125 19:54:59.380346 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.370967 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.375749 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.377764 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.379839 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.395929 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.415114 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521644 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521697 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521718 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521744 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521771 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521786 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521811 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521838 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521853 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521871 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.521981 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522023 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522044 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522093 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjh82\" (UniqueName: \"kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522167 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522194 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522224 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522244 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522260 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522281 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522333 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522370 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkqrb\" (UniqueName: \"kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522390 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522406 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522428 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522474 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522503 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.522557 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624279 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624472 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624503 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624637 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624709 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624788 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjh82\" (UniqueName: \"kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624580 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624915 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.624982 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625056 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625124 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625212 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625288 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625365 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625417 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625003 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625322 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625510 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkqrb\" (UniqueName: \"kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625543 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625065 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625597 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625397 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625652 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625686 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625743 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625753 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625808 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625856 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625901 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.625954 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626002 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626061 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626101 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626165 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626204 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626214 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626222 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626313 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626414 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626465 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626507 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626167 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626643 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626665 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626732 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626872 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626932 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626988 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.626949 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.627130 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.632347 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.633745 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.637361 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.639710 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.645123 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjh82\" (UniqueName: \"kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.647018 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkqrb\" (UniqueName: \"kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.647477 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.652095 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.653823 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.662541 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.727538 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:02 crc kubenswrapper[4759]: I1125 19:55:02.734966 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:03 crc kubenswrapper[4759]: W1125 19:55:03.154014 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbff8231_e04f_404a_9cb1_dc4d03570b3c.slice/crio-169316977ba2816caae2ac87d2aa12a78dc4a981ae1f51cc90de6d26563fae16 WatchSource:0}: Error finding container 169316977ba2816caae2ac87d2aa12a78dc4a981ae1f51cc90de6d26563fae16: Status 404 returned error can't find the container with id 169316977ba2816caae2ac87d2aa12a78dc4a981ae1f51cc90de6d26563fae16 Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.154027 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.203760 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:55:03 crc kubenswrapper[4759]: W1125 19:55:03.210283 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61780c52_a2da_492a_8107_3f704999e2af.slice/crio-bf7a08ffcec93abbf8f89bc0f36be349d3ef01613abd12857a48b34f2b7fcd83 WatchSource:0}: Error finding container bf7a08ffcec93abbf8f89bc0f36be349d3ef01613abd12857a48b34f2b7fcd83: Status 404 returned error can't find the container with id bf7a08ffcec93abbf8f89bc0f36be349d3ef01613abd12857a48b34f2b7fcd83 Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.390271 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerStarted","Data":"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466"} Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.390593 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerStarted","Data":"169316977ba2816caae2ac87d2aa12a78dc4a981ae1f51cc90de6d26563fae16"} Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.391691 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerStarted","Data":"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9"} Nov 25 19:55:03 crc kubenswrapper[4759]: I1125 19:55:03.391734 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerStarted","Data":"bf7a08ffcec93abbf8f89bc0f36be349d3ef01613abd12857a48b34f2b7fcd83"} Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.406387 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerStarted","Data":"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325"} Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.406704 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerStarted","Data":"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a"} Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.410500 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerStarted","Data":"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d"} Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.410604 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerStarted","Data":"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd"} Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.441900 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.441869037 podStartE2EDuration="3.441869037s" podCreationTimestamp="2025-11-25 19:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:55:04.432578522 +0000 UTC m=+1295.018526293" watchObservedRunningTime="2025-11-25 19:55:04.441869037 +0000 UTC m=+1295.027816778" Nov 25 19:55:04 crc kubenswrapper[4759]: I1125 19:55:04.458842 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.458822092 podStartE2EDuration="3.458822092s" podCreationTimestamp="2025-11-25 19:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 19:55:04.457274 +0000 UTC m=+1295.043221781" watchObservedRunningTime="2025-11-25 19:55:04.458822092 +0000 UTC m=+1295.044769803" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.728537 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.730572 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.730593 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.735171 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.735220 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.735231 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.764252 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.764665 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.764984 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.766345 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.776263 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:12 crc kubenswrapper[4759]: I1125 19:55:12.778082 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.498809 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.499139 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.499152 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.499164 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.499173 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.499183 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.509724 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.510037 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.510156 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.511052 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.512910 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:13 crc kubenswrapper[4759]: I1125 19:55:13.517797 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:55:19 crc kubenswrapper[4759]: I1125 19:55:19.496328 4759 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-zlfnc" podUID="f4b24bd8-f201-4cbb-8a6c-1821e1c1002c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 25 19:56:54 crc kubenswrapper[4759]: I1125 19:56:54.896120 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:56:54 crc kubenswrapper[4759]: I1125 19:56:54.896891 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:57:24 crc kubenswrapper[4759]: I1125 19:57:24.895994 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:57:24 crc kubenswrapper[4759]: I1125 19:57:24.896539 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:57:54 crc kubenswrapper[4759]: I1125 19:57:54.896020 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 19:57:54 crc kubenswrapper[4759]: I1125 19:57:54.896561 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 19:57:54 crc kubenswrapper[4759]: I1125 19:57:54.896601 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 19:57:54 crc kubenswrapper[4759]: I1125 19:57:54.897115 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 19:57:54 crc kubenswrapper[4759]: I1125 19:57:54.897232 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2" gracePeriod=600 Nov 25 19:57:55 crc kubenswrapper[4759]: I1125 19:57:55.835513 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2" exitCode=0 Nov 25 19:57:55 crc kubenswrapper[4759]: I1125 19:57:55.835595 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2"} Nov 25 19:57:55 crc kubenswrapper[4759]: I1125 19:57:55.835805 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad"} Nov 25 19:57:55 crc kubenswrapper[4759]: I1125 19:57:55.835828 4759 scope.go:117] "RemoveContainer" containerID="ca3cac62c6e8022263aee3c67ef8ba67c5aa2d87ca61b8e67e5f101bf8e31930" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.061631 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.062647 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-log" containerID="cri-o://02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.062684 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-api" containerID="cri-o://aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.062734 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-httpd" containerID="cri-o://7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.079979 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.080271 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-log" containerID="cri-o://727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.080558 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-api" containerID="cri-o://0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.080569 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-httpd" containerID="cri-o://a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a" gracePeriod=30 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.387692 4759 generic.go:334] "Generic (PLEG): container finished" podID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerID="a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a" exitCode=0 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.388050 4759 generic.go:334] "Generic (PLEG): container finished" podID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerID="727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466" exitCode=143 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.387773 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerDied","Data":"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a"} Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.388134 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerDied","Data":"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466"} Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.390565 4759 generic.go:334] "Generic (PLEG): container finished" podID="61780c52-a2da-492a-8107-3f704999e2af" containerID="7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd" exitCode=0 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.390589 4759 generic.go:334] "Generic (PLEG): container finished" podID="61780c52-a2da-492a-8107-3f704999e2af" containerID="02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9" exitCode=143 Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.390608 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerDied","Data":"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd"} Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.390630 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerDied","Data":"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9"} Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.865711 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.869798 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949770 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949822 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949841 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949858 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949896 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjh82\" (UniqueName: \"kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949927 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949951 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949969 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.949991 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950011 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950026 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950050 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950068 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950090 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950108 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950123 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950136 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950159 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkqrb\" (UniqueName: \"kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950176 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950218 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950235 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950254 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950265 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950302 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950314 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950334 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950352 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules\") pod \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\" (UID: \"bbff8231-e04f-404a-9cb1-dc4d03570b3c\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950373 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick\") pod \"61780c52-a2da-492a-8107-3f704999e2af\" (UID: \"61780c52-a2da-492a-8107-3f704999e2af\") " Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.950707 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.951427 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.957628 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.957893 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.957920 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev" (OuterVolumeSpecName: "dev") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.963401 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys" (OuterVolumeSpecName: "sys") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.964435 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs" (OuterVolumeSpecName: "logs") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.964499 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.964526 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.964994 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.965066 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys" (OuterVolumeSpecName: "sys") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.965090 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.965110 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run" (OuterVolumeSpecName: "run") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.966776 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs" (OuterVolumeSpecName: "logs") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.966816 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.966899 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.967573 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.968742 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev" (OuterVolumeSpecName: "dev") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.969835 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.969908 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run" (OuterVolumeSpecName: "run") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.975616 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts" (OuterVolumeSpecName: "scripts") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.975644 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb" (OuterVolumeSpecName: "kube-api-access-qkqrb") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "kube-api-access-qkqrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.979653 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.979666 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82" (OuterVolumeSpecName: "kube-api-access-xjh82") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "kube-api-access-xjh82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:59:04 crc kubenswrapper[4759]: I1125 19:59:04.986714 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.014630 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts" (OuterVolumeSpecName: "scripts") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051772 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051806 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051817 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-sys\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051826 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051835 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051854 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051864 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051875 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051889 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051898 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051906 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051914 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjh82\" (UniqueName: \"kubernetes.io/projected/61780c52-a2da-492a-8107-3f704999e2af-kube-api-access-xjh82\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051922 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbff8231-e04f-404a-9cb1-dc4d03570b3c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051930 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051938 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-logs\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051946 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051953 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051960 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051971 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051979 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051988 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/61780c52-a2da-492a-8107-3f704999e2af-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.051996 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61780c52-a2da-492a-8107-3f704999e2af-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.052006 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bbff8231-e04f-404a-9cb1-dc4d03570b3c-dev\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.052018 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.052026 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkqrb\" (UniqueName: \"kubernetes.io/projected/bbff8231-e04f-404a-9cb1-dc4d03570b3c-kube-api-access-qkqrb\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.052033 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.098384 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.099772 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.102989 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.105243 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.105507 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data" (OuterVolumeSpecName: "config-data") pod "61780c52-a2da-492a-8107-3f704999e2af" (UID: "61780c52-a2da-492a-8107-3f704999e2af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.132718 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data" (OuterVolumeSpecName: "config-data") pod "bbff8231-e04f-404a-9cb1-dc4d03570b3c" (UID: "bbff8231-e04f-404a-9cb1-dc4d03570b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153350 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153381 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153391 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbff8231-e04f-404a-9cb1-dc4d03570b3c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153405 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153456 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.153465 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61780c52-a2da-492a-8107-3f704999e2af-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.399477 4759 generic.go:334] "Generic (PLEG): container finished" podID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerID="0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325" exitCode=0 Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.399562 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerDied","Data":"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325"} Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.399594 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bbff8231-e04f-404a-9cb1-dc4d03570b3c","Type":"ContainerDied","Data":"169316977ba2816caae2ac87d2aa12a78dc4a981ae1f51cc90de6d26563fae16"} Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.399616 4759 scope.go:117] "RemoveContainer" containerID="0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.399972 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.403222 4759 generic.go:334] "Generic (PLEG): container finished" podID="61780c52-a2da-492a-8107-3f704999e2af" containerID="aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d" exitCode=0 Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.403270 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerDied","Data":"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d"} Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.403315 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"61780c52-a2da-492a-8107-3f704999e2af","Type":"ContainerDied","Data":"bf7a08ffcec93abbf8f89bc0f36be349d3ef01613abd12857a48b34f2b7fcd83"} Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.403275 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.418914 4759 scope.go:117] "RemoveContainer" containerID="a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.444258 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.452709 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.456682 4759 scope.go:117] "RemoveContainer" containerID="727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.466823 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.471584 4759 scope.go:117] "RemoveContainer" containerID="0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.472052 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325\": container with ID starting with 0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325 not found: ID does not exist" containerID="0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.472095 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325"} err="failed to get container status \"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325\": rpc error: code = NotFound desc = could not find container \"0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325\": container with ID starting with 0cd4de8aa1098ccb47a69c9698e4c15d9b685bfd3be691c8c722be28397e1325 not found: ID does not exist" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.472120 4759 scope.go:117] "RemoveContainer" containerID="a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.472639 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a\": container with ID starting with a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a not found: ID does not exist" containerID="a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.472711 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a"} err="failed to get container status \"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a\": rpc error: code = NotFound desc = could not find container \"a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a\": container with ID starting with a4a28c3f63c97b23a418d7cbc5f1b3008b0bbc64cdd5aac8cddb9c8797ce474a not found: ID does not exist" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.472751 4759 scope.go:117] "RemoveContainer" containerID="727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.473148 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466\": container with ID starting with 727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466 not found: ID does not exist" containerID="727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.473179 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466"} err="failed to get container status \"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466\": rpc error: code = NotFound desc = could not find container \"727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466\": container with ID starting with 727fe81dc2b4c1b9d4d39d53010c91aaaffd8f0b8fa74d43a959418785577466 not found: ID does not exist" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.473196 4759 scope.go:117] "RemoveContainer" containerID="aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.475542 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.492081 4759 scope.go:117] "RemoveContainer" containerID="7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.511105 4759 scope.go:117] "RemoveContainer" containerID="02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.528100 4759 scope.go:117] "RemoveContainer" containerID="aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.528561 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d\": container with ID starting with aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d not found: ID does not exist" containerID="aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.528599 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d"} err="failed to get container status \"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d\": rpc error: code = NotFound desc = could not find container \"aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d\": container with ID starting with aa45805bec9d0c1ffc2f814dbdce16ff891a95c1b32b2015ad8c0d57624b290d not found: ID does not exist" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.528630 4759 scope.go:117] "RemoveContainer" containerID="7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.528998 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd\": container with ID starting with 7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd not found: ID does not exist" containerID="7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.529037 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd"} err="failed to get container status \"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd\": rpc error: code = NotFound desc = could not find container \"7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd\": container with ID starting with 7dd0e59b9f514b038e42d0d3504aba3c8216229c26c85edeaf1c28efc57fbdfd not found: ID does not exist" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.529067 4759 scope.go:117] "RemoveContainer" containerID="02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9" Nov 25 19:59:05 crc kubenswrapper[4759]: E1125 19:59:05.529490 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9\": container with ID starting with 02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9 not found: ID does not exist" containerID="02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9" Nov 25 19:59:05 crc kubenswrapper[4759]: I1125 19:59:05.529535 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9"} err="failed to get container status \"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9\": rpc error: code = NotFound desc = could not find container \"02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9\": container with ID starting with 02623ca2100df6331a13fff0d23775a49776c750bbef27b6a59cfb00d5aeadd9 not found: ID does not exist" Nov 25 19:59:06 crc kubenswrapper[4759]: I1125 19:59:06.117370 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61780c52-a2da-492a-8107-3f704999e2af" path="/var/lib/kubelet/pods/61780c52-a2da-492a-8107-3f704999e2af/volumes" Nov 25 19:59:06 crc kubenswrapper[4759]: I1125 19:59:06.118622 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" path="/var/lib/kubelet/pods/bbff8231-e04f-404a-9cb1-dc4d03570b3c/volumes" Nov 25 19:59:30 crc kubenswrapper[4759]: I1125 19:59:30.778597 4759 scope.go:117] "RemoveContainer" containerID="e37d654761b2b5e0ae774004dcd1fee6685d57e31e272fced6511fe3a5de485c" Nov 25 19:59:30 crc kubenswrapper[4759]: I1125 19:59:30.805642 4759 scope.go:117] "RemoveContainer" containerID="6f6ce9372ba72e277d6ff0bd00522e4e3f755795987065d119db9f479d322af6" Nov 25 19:59:30 crc kubenswrapper[4759]: I1125 19:59:30.858211 4759 scope.go:117] "RemoveContainer" containerID="ff55e90629e22372aebdf02d0242a6aee9ea7beda4c51ac0654b12807c33b920" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.036467 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037077 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037090 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037115 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037121 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037134 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037142 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037162 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037168 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037184 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037190 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: E1125 19:59:33.037198 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037203 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037323 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037334 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037342 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-log" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037350 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbff8231-e04f-404a-9cb1-dc4d03570b3c" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037359 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-httpd" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.037370 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="61780c52-a2da-492a-8107-3f704999e2af" containerName="glance-api" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.038413 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.051992 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.193022 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqcwh\" (UniqueName: \"kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.193182 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.193214 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.295188 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.295305 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqcwh\" (UniqueName: \"kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.296213 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.296271 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.296574 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.329535 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqcwh\" (UniqueName: \"kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh\") pod \"community-operators-jjnqk\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.398181 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:33 crc kubenswrapper[4759]: I1125 19:59:33.684658 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:34 crc kubenswrapper[4759]: I1125 19:59:34.648806 4759 generic.go:334] "Generic (PLEG): container finished" podID="cb7e9783-0635-4465-855b-84ef2825d185" containerID="22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd" exitCode=0 Nov 25 19:59:34 crc kubenswrapper[4759]: I1125 19:59:34.648849 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerDied","Data":"22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd"} Nov 25 19:59:34 crc kubenswrapper[4759]: I1125 19:59:34.649107 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerStarted","Data":"d664e7f50605dfcd3da599ea2135f826b1240178705aa2c748e8f497581a958e"} Nov 25 19:59:34 crc kubenswrapper[4759]: I1125 19:59:34.650911 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 19:59:36 crc kubenswrapper[4759]: I1125 19:59:36.666927 4759 generic.go:334] "Generic (PLEG): container finished" podID="cb7e9783-0635-4465-855b-84ef2825d185" containerID="b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34" exitCode=0 Nov 25 19:59:36 crc kubenswrapper[4759]: I1125 19:59:36.667038 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerDied","Data":"b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34"} Nov 25 19:59:37 crc kubenswrapper[4759]: I1125 19:59:37.676941 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerStarted","Data":"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc"} Nov 25 19:59:37 crc kubenswrapper[4759]: I1125 19:59:37.703471 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jjnqk" podStartSLOduration=2.223119259 podStartE2EDuration="4.703436738s" podCreationTimestamp="2025-11-25 19:59:33 +0000 UTC" firstStartedPulling="2025-11-25 19:59:34.65068259 +0000 UTC m=+1565.236630291" lastFinishedPulling="2025-11-25 19:59:37.131000069 +0000 UTC m=+1567.716947770" observedRunningTime="2025-11-25 19:59:37.701385342 +0000 UTC m=+1568.287333053" watchObservedRunningTime="2025-11-25 19:59:37.703436738 +0000 UTC m=+1568.289384439" Nov 25 19:59:43 crc kubenswrapper[4759]: I1125 19:59:43.399285 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:43 crc kubenswrapper[4759]: I1125 19:59:43.400008 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:43 crc kubenswrapper[4759]: I1125 19:59:43.442445 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:43 crc kubenswrapper[4759]: I1125 19:59:43.760094 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:43 crc kubenswrapper[4759]: I1125 19:59:43.804434 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:45 crc kubenswrapper[4759]: I1125 19:59:45.738249 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jjnqk" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="registry-server" containerID="cri-o://ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc" gracePeriod=2 Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.124927 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.234028 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqcwh\" (UniqueName: \"kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh\") pod \"cb7e9783-0635-4465-855b-84ef2825d185\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.234143 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities\") pod \"cb7e9783-0635-4465-855b-84ef2825d185\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.234201 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content\") pod \"cb7e9783-0635-4465-855b-84ef2825d185\" (UID: \"cb7e9783-0635-4465-855b-84ef2825d185\") " Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.237006 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities" (OuterVolumeSpecName: "utilities") pod "cb7e9783-0635-4465-855b-84ef2825d185" (UID: "cb7e9783-0635-4465-855b-84ef2825d185"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.240272 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh" (OuterVolumeSpecName: "kube-api-access-cqcwh") pod "cb7e9783-0635-4465-855b-84ef2825d185" (UID: "cb7e9783-0635-4465-855b-84ef2825d185"). InnerVolumeSpecName "kube-api-access-cqcwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.335926 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.335960 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqcwh\" (UniqueName: \"kubernetes.io/projected/cb7e9783-0635-4465-855b-84ef2825d185-kube-api-access-cqcwh\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.477523 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb7e9783-0635-4465-855b-84ef2825d185" (UID: "cb7e9783-0635-4465-855b-84ef2825d185"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.538764 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb7e9783-0635-4465-855b-84ef2825d185-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.747166 4759 generic.go:334] "Generic (PLEG): container finished" podID="cb7e9783-0635-4465-855b-84ef2825d185" containerID="ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc" exitCode=0 Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.747227 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerDied","Data":"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc"} Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.747268 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjnqk" event={"ID":"cb7e9783-0635-4465-855b-84ef2825d185","Type":"ContainerDied","Data":"d664e7f50605dfcd3da599ea2135f826b1240178705aa2c748e8f497581a958e"} Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.747301 4759 scope.go:117] "RemoveContainer" containerID="ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.747295 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjnqk" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.764437 4759 scope.go:117] "RemoveContainer" containerID="b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.781580 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.788167 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jjnqk"] Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.797110 4759 scope.go:117] "RemoveContainer" containerID="22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.811531 4759 scope.go:117] "RemoveContainer" containerID="ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc" Nov 25 19:59:46 crc kubenswrapper[4759]: E1125 19:59:46.812021 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc\": container with ID starting with ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc not found: ID does not exist" containerID="ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.812089 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc"} err="failed to get container status \"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc\": rpc error: code = NotFound desc = could not find container \"ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc\": container with ID starting with ede8e481f0bce7e2c5661c8c5abdb251d32546adad107aea122e521bda37d4fc not found: ID does not exist" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.812128 4759 scope.go:117] "RemoveContainer" containerID="b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34" Nov 25 19:59:46 crc kubenswrapper[4759]: E1125 19:59:46.812575 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34\": container with ID starting with b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34 not found: ID does not exist" containerID="b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.812693 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34"} err="failed to get container status \"b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34\": rpc error: code = NotFound desc = could not find container \"b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34\": container with ID starting with b4f5e1db142950caa90f5f0704a234952490d344e383c909475e0c0975e52f34 not found: ID does not exist" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.812784 4759 scope.go:117] "RemoveContainer" containerID="22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd" Nov 25 19:59:46 crc kubenswrapper[4759]: E1125 19:59:46.813163 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd\": container with ID starting with 22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd not found: ID does not exist" containerID="22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd" Nov 25 19:59:46 crc kubenswrapper[4759]: I1125 19:59:46.813192 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd"} err="failed to get container status \"22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd\": rpc error: code = NotFound desc = could not find container \"22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd\": container with ID starting with 22a4c418033206fe89ca82c62d7164e2b84db94bf1c02ad38e591cdea6df06dd not found: ID does not exist" Nov 25 19:59:48 crc kubenswrapper[4759]: I1125 19:59:48.121981 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb7e9783-0635-4465-855b-84ef2825d185" path="/var/lib/kubelet/pods/cb7e9783-0635-4465-855b-84ef2825d185/volumes" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.156806 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr"] Nov 25 20:00:00 crc kubenswrapper[4759]: E1125 20:00:00.157778 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="extract-utilities" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.157798 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="extract-utilities" Nov 25 20:00:00 crc kubenswrapper[4759]: E1125 20:00:00.157827 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="registry-server" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.157835 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="registry-server" Nov 25 20:00:00 crc kubenswrapper[4759]: E1125 20:00:00.157856 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="extract-content" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.157864 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="extract-content" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.158012 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb7e9783-0635-4465-855b-84ef2825d185" containerName="registry-server" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.158665 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.165988 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.167340 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.169058 4759 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.172984 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.177602 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.178579 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.191512 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.216586 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.237784 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269603 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269647 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269703 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269766 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22nzl\" (UniqueName: \"kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269790 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269809 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269880 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vlw6\" (UniqueName: \"kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269910 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvf7\" (UniqueName: \"kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.269965 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.295473 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.308925 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371131 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22nzl\" (UniqueName: \"kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371182 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371207 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371256 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vlw6\" (UniqueName: \"kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371290 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvf7\" (UniqueName: \"kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371323 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.371390 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.372485 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.375651 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.375707 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.375919 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.388859 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vlw6\" (UniqueName: \"kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940168dsllr\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.389251 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvf7\" (UniqueName: \"kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.391611 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22nzl\" (UniqueName: \"kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl\") pod \"collect-profiles-29401680-grqsr\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.484244 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.497827 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.508907 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.897555 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.966461 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr"] Nov 25 20:00:00 crc kubenswrapper[4759]: I1125 20:00:00.978135 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67"] Nov 25 20:00:00 crc kubenswrapper[4759]: W1125 20:00:00.978569 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0abf4b9b_8f46_4cd7_8c13_f0c6dc1cd142.slice/crio-563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc WatchSource:0}: Error finding container 563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc: Status 404 returned error can't find the container with id 563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc Nov 25 20:00:00 crc kubenswrapper[4759]: W1125 20:00:00.979692 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc61252ae_a6de_47a2_9355_d9392583883c.slice/crio-8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966 WatchSource:0}: Error finding container 8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966: Status 404 returned error can't find the container with id 8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966 Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.869387 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" event={"ID":"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142","Type":"ContainerStarted","Data":"dd563ef69d90ec4d702b8650a3d976c0dd6b6061ee8ac82425ad4bab117a1ffa"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.869990 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" event={"ID":"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142","Type":"ContainerStarted","Data":"563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.876022 4759 generic.go:334] "Generic (PLEG): container finished" podID="35304ddf-a39f-40f0-bb39-8b9e9c168731" containerID="f668d26f49f0a2400c0439d76724ac721e5575842dda3776820be95c4596b3b1" exitCode=0 Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.876144 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" event={"ID":"35304ddf-a39f-40f0-bb39-8b9e9c168731","Type":"ContainerDied","Data":"f668d26f49f0a2400c0439d76724ac721e5575842dda3776820be95c4596b3b1"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.876173 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" event={"ID":"35304ddf-a39f-40f0-bb39-8b9e9c168731","Type":"ContainerStarted","Data":"5e8d17ecf1eaf0b1c26c0637c848f0a756dbd22777e011a85a84da9568aede3d"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.881932 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" event={"ID":"c61252ae-a6de-47a2-9355-d9392583883c","Type":"ContainerStarted","Data":"9b5c8ec545cdd11486c9d79b81e0f8e3ed9dca45e52b6179129e140f3fd8cea1"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.881984 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" event={"ID":"c61252ae-a6de-47a2-9355-d9392583883c","Type":"ContainerStarted","Data":"8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966"} Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.900890 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" podStartSLOduration=1.900869178 podStartE2EDuration="1.900869178s" podCreationTimestamp="2025-11-25 20:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:00:01.891563653 +0000 UTC m=+1592.477511354" watchObservedRunningTime="2025-11-25 20:00:01.900869178 +0000 UTC m=+1592.486816889" Nov 25 20:00:01 crc kubenswrapper[4759]: I1125 20:00:01.937675 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" podStartSLOduration=1.93765254 podStartE2EDuration="1.93765254s" podCreationTimestamp="2025-11-25 20:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:00:01.933490676 +0000 UTC m=+1592.519438387" watchObservedRunningTime="2025-11-25 20:00:01.93765254 +0000 UTC m=+1592.523600241" Nov 25 20:00:02 crc kubenswrapper[4759]: I1125 20:00:02.892624 4759 generic.go:334] "Generic (PLEG): container finished" podID="0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" containerID="dd563ef69d90ec4d702b8650a3d976c0dd6b6061ee8ac82425ad4bab117a1ffa" exitCode=0 Nov 25 20:00:02 crc kubenswrapper[4759]: I1125 20:00:02.892705 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" event={"ID":"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142","Type":"ContainerDied","Data":"dd563ef69d90ec4d702b8650a3d976c0dd6b6061ee8ac82425ad4bab117a1ffa"} Nov 25 20:00:02 crc kubenswrapper[4759]: I1125 20:00:02.895158 4759 generic.go:334] "Generic (PLEG): container finished" podID="c61252ae-a6de-47a2-9355-d9392583883c" containerID="9b5c8ec545cdd11486c9d79b81e0f8e3ed9dca45e52b6179129e140f3fd8cea1" exitCode=0 Nov 25 20:00:02 crc kubenswrapper[4759]: I1125 20:00:02.895237 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" event={"ID":"c61252ae-a6de-47a2-9355-d9392583883c","Type":"ContainerDied","Data":"9b5c8ec545cdd11486c9d79b81e0f8e3ed9dca45e52b6179129e140f3fd8cea1"} Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.176760 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.319890 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume\") pod \"35304ddf-a39f-40f0-bb39-8b9e9c168731\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.320430 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume" (OuterVolumeSpecName: "config-volume") pod "35304ddf-a39f-40f0-bb39-8b9e9c168731" (UID: "35304ddf-a39f-40f0-bb39-8b9e9c168731"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.320607 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume\") pod \"35304ddf-a39f-40f0-bb39-8b9e9c168731\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.320775 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22nzl\" (UniqueName: \"kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl\") pod \"35304ddf-a39f-40f0-bb39-8b9e9c168731\" (UID: \"35304ddf-a39f-40f0-bb39-8b9e9c168731\") " Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.321484 4759 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35304ddf-a39f-40f0-bb39-8b9e9c168731-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.327820 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "35304ddf-a39f-40f0-bb39-8b9e9c168731" (UID: "35304ddf-a39f-40f0-bb39-8b9e9c168731"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.328081 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl" (OuterVolumeSpecName: "kube-api-access-22nzl") pod "35304ddf-a39f-40f0-bb39-8b9e9c168731" (UID: "35304ddf-a39f-40f0-bb39-8b9e9c168731"). InnerVolumeSpecName "kube-api-access-22nzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.423289 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22nzl\" (UniqueName: \"kubernetes.io/projected/35304ddf-a39f-40f0-bb39-8b9e9c168731-kube-api-access-22nzl\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.423355 4759 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35304ddf-a39f-40f0-bb39-8b9e9c168731-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.909163 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" event={"ID":"35304ddf-a39f-40f0-bb39-8b9e9c168731","Type":"ContainerDied","Data":"5e8d17ecf1eaf0b1c26c0637c848f0a756dbd22777e011a85a84da9568aede3d"} Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.910739 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e8d17ecf1eaf0b1c26c0637c848f0a756dbd22777e011a85a84da9568aede3d" Nov 25 20:00:03 crc kubenswrapper[4759]: I1125 20:00:03.909286 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401680-grqsr" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.243722 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.249638 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.336671 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vlw6\" (UniqueName: \"kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6\") pod \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.336772 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data\") pod \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.336848 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data\") pod \"c61252ae-a6de-47a2-9355-d9392583883c\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.336880 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwvf7\" (UniqueName: \"kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7\") pod \"c61252ae-a6de-47a2-9355-d9392583883c\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.337011 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"c61252ae-a6de-47a2-9355-d9392583883c\" (UID: \"c61252ae-a6de-47a2-9355-d9392583883c\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.337040 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\" (UID: \"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142\") " Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.342796 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" (UID: "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.342798 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" (UID: "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.342839 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7" (OuterVolumeSpecName: "kube-api-access-zwvf7") pod "c61252ae-a6de-47a2-9355-d9392583883c" (UID: "c61252ae-a6de-47a2-9355-d9392583883c"). InnerVolumeSpecName "kube-api-access-zwvf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.342948 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "c61252ae-a6de-47a2-9355-d9392583883c" (UID: "c61252ae-a6de-47a2-9355-d9392583883c"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.344003 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6" (OuterVolumeSpecName: "kube-api-access-7vlw6") pod "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" (UID: "0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142"). InnerVolumeSpecName "kube-api-access-7vlw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.348386 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "c61252ae-a6de-47a2-9355-d9392583883c" (UID: "c61252ae-a6de-47a2-9355-d9392583883c"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.438401 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vlw6\" (UniqueName: \"kubernetes.io/projected/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-kube-api-access-7vlw6\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.438706 4759 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.438778 4759 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/c61252ae-a6de-47a2-9355-d9392583883c-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.438834 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwvf7\" (UniqueName: \"kubernetes.io/projected/c61252ae-a6de-47a2-9355-d9392583883c-kube-api-access-zwvf7\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.915846 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.915845 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67" event={"ID":"c61252ae-a6de-47a2-9355-d9392583883c","Type":"ContainerDied","Data":"8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966"} Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.916010 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8307a7d979eb35f02ac71710cee507876bf5f2e75cf2a5e98214b0b396f40966" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.916825 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" event={"ID":"0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142","Type":"ContainerDied","Data":"563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc"} Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.916869 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="563e0d7e3d00b4ad849c2e91436ebca4ac7389a8741828ef17fd569b7c8a93fc" Nov 25 20:00:04 crc kubenswrapper[4759]: I1125 20:00:04.916899 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.125305 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:16 crc kubenswrapper[4759]: E1125 20:00:16.126310 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35304ddf-a39f-40f0-bb39-8b9e9c168731" containerName="collect-profiles" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126331 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="35304ddf-a39f-40f0-bb39-8b9e9c168731" containerName="collect-profiles" Nov 25 20:00:16 crc kubenswrapper[4759]: E1125 20:00:16.126356 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126371 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: E1125 20:00:16.126408 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61252ae-a6de-47a2-9355-d9392583883c" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126420 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61252ae-a6de-47a2-9355-d9392583883c" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126678 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="35304ddf-a39f-40f0-bb39-8b9e9c168731" containerName="collect-profiles" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126707 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.126733 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61252ae-a6de-47a2-9355-d9392583883c" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.130163 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.130344 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.244215 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.245193 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.245474 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9jgs\" (UniqueName: \"kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.347647 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.347713 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9jgs\" (UniqueName: \"kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.347771 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.348242 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.348703 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.373992 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9jgs\" (UniqueName: \"kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs\") pod \"redhat-operators-qr2kd\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.466041 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:16 crc kubenswrapper[4759]: I1125 20:00:16.886813 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:17 crc kubenswrapper[4759]: I1125 20:00:17.030926 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerStarted","Data":"7811e44fbf44460013f4a4861c21ed5b434a1c68f744b65a516c9e194e409934"} Nov 25 20:00:18 crc kubenswrapper[4759]: I1125 20:00:18.042097 4759 generic.go:334] "Generic (PLEG): container finished" podID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerID="9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d" exitCode=0 Nov 25 20:00:18 crc kubenswrapper[4759]: I1125 20:00:18.042177 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerDied","Data":"9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d"} Nov 25 20:00:19 crc kubenswrapper[4759]: I1125 20:00:19.050424 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerStarted","Data":"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375"} Nov 25 20:00:20 crc kubenswrapper[4759]: I1125 20:00:20.058275 4759 generic.go:334] "Generic (PLEG): container finished" podID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerID="ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375" exitCode=0 Nov 25 20:00:20 crc kubenswrapper[4759]: I1125 20:00:20.058319 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerDied","Data":"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375"} Nov 25 20:00:21 crc kubenswrapper[4759]: I1125 20:00:21.066640 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerStarted","Data":"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d"} Nov 25 20:00:21 crc kubenswrapper[4759]: I1125 20:00:21.090348 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qr2kd" podStartSLOduration=2.660349684 podStartE2EDuration="5.090328439s" podCreationTimestamp="2025-11-25 20:00:16 +0000 UTC" firstStartedPulling="2025-11-25 20:00:18.049279382 +0000 UTC m=+1608.635227123" lastFinishedPulling="2025-11-25 20:00:20.479258167 +0000 UTC m=+1611.065205878" observedRunningTime="2025-11-25 20:00:21.087021889 +0000 UTC m=+1611.672969590" watchObservedRunningTime="2025-11-25 20:00:21.090328439 +0000 UTC m=+1611.676276140" Nov 25 20:00:24 crc kubenswrapper[4759]: I1125 20:00:24.895703 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:00:24 crc kubenswrapper[4759]: I1125 20:00:24.896278 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:00:26 crc kubenswrapper[4759]: I1125 20:00:26.466312 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:26 crc kubenswrapper[4759]: I1125 20:00:26.466617 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:26 crc kubenswrapper[4759]: I1125 20:00:26.533725 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:27 crc kubenswrapper[4759]: I1125 20:00:27.179613 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:27 crc kubenswrapper[4759]: I1125 20:00:27.234923 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:29 crc kubenswrapper[4759]: I1125 20:00:29.130679 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qr2kd" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="registry-server" containerID="cri-o://f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d" gracePeriod=2 Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.049341 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.141402 4759 generic.go:334] "Generic (PLEG): container finished" podID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerID="f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d" exitCode=0 Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.141455 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerDied","Data":"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d"} Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.141469 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr2kd" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.141483 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr2kd" event={"ID":"45828622-20a2-4839-ab4a-acbe8b3a02e6","Type":"ContainerDied","Data":"7811e44fbf44460013f4a4861c21ed5b434a1c68f744b65a516c9e194e409934"} Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.141499 4759 scope.go:117] "RemoveContainer" containerID="f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.161431 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities\") pod \"45828622-20a2-4839-ab4a-acbe8b3a02e6\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.161527 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content\") pod \"45828622-20a2-4839-ab4a-acbe8b3a02e6\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.161626 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9jgs\" (UniqueName: \"kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs\") pod \"45828622-20a2-4839-ab4a-acbe8b3a02e6\" (UID: \"45828622-20a2-4839-ab4a-acbe8b3a02e6\") " Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.162939 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities" (OuterVolumeSpecName: "utilities") pod "45828622-20a2-4839-ab4a-acbe8b3a02e6" (UID: "45828622-20a2-4839-ab4a-acbe8b3a02e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.166067 4759 scope.go:117] "RemoveContainer" containerID="ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.169791 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs" (OuterVolumeSpecName: "kube-api-access-v9jgs") pod "45828622-20a2-4839-ab4a-acbe8b3a02e6" (UID: "45828622-20a2-4839-ab4a-acbe8b3a02e6"). InnerVolumeSpecName "kube-api-access-v9jgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.208080 4759 scope.go:117] "RemoveContainer" containerID="9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.236675 4759 scope.go:117] "RemoveContainer" containerID="f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d" Nov 25 20:00:30 crc kubenswrapper[4759]: E1125 20:00:30.237102 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d\": container with ID starting with f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d not found: ID does not exist" containerID="f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.237136 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d"} err="failed to get container status \"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d\": rpc error: code = NotFound desc = could not find container \"f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d\": container with ID starting with f747b67669930a3293d7e39d256a19ae412aff23bb874f094519bd772d49910d not found: ID does not exist" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.237155 4759 scope.go:117] "RemoveContainer" containerID="ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375" Nov 25 20:00:30 crc kubenswrapper[4759]: E1125 20:00:30.237381 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375\": container with ID starting with ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375 not found: ID does not exist" containerID="ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.237403 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375"} err="failed to get container status \"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375\": rpc error: code = NotFound desc = could not find container \"ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375\": container with ID starting with ed6acea1abe9827a31ea019fce3562f749f5c9ee87e4414b059e2beba3fda375 not found: ID does not exist" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.237419 4759 scope.go:117] "RemoveContainer" containerID="9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d" Nov 25 20:00:30 crc kubenswrapper[4759]: E1125 20:00:30.237703 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d\": container with ID starting with 9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d not found: ID does not exist" containerID="9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.237733 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d"} err="failed to get container status \"9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d\": rpc error: code = NotFound desc = could not find container \"9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d\": container with ID starting with 9c9898a21f4bf474b119f12b300e8be8c1b79e663e8204f7c91ef57127ec211d not found: ID does not exist" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.264035 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9jgs\" (UniqueName: \"kubernetes.io/projected/45828622-20a2-4839-ab4a-acbe8b3a02e6-kube-api-access-v9jgs\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.264077 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.908195 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45828622-20a2-4839-ab4a-acbe8b3a02e6" (UID: "45828622-20a2-4839-ab4a-acbe8b3a02e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.955217 4759 scope.go:117] "RemoveContainer" containerID="f5b621c8a24752948a9a7e746ea52014f05c59397538038fcc2e18aaa9d3193a" Nov 25 20:00:30 crc kubenswrapper[4759]: I1125 20:00:30.973553 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45828622-20a2-4839-ab4a-acbe8b3a02e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.010862 4759 scope.go:117] "RemoveContainer" containerID="6cab315d8ded72ac3a7d990022f75fefa10f1c422c74046180f779666a9e742f" Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.041246 4759 scope.go:117] "RemoveContainer" containerID="c051c659087c6b9bcb1fbd7a9b36f64920c5039e1b9b595eb0fa4c47c400cc97" Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.066915 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr"] Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.080160 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hc6nr"] Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.086801 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-ef97-account-create-update-4bbpr"] Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.090896 4759 scope.go:117] "RemoveContainer" containerID="9e7ab002ae6e5148facfe93c5b35ebacb9331334114aa2e9059f7a1abbe5ae1f" Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.093508 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-hc6nr"] Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.127019 4759 scope.go:117] "RemoveContainer" containerID="c4554495742a389f8c8b9f0f25ae299ac86225f7b95f8111af1249defbda7acd" Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.181888 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:31 crc kubenswrapper[4759]: I1125 20:00:31.189274 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qr2kd"] Nov 25 20:00:32 crc kubenswrapper[4759]: I1125 20:00:32.119670 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" path="/var/lib/kubelet/pods/45828622-20a2-4839-ab4a-acbe8b3a02e6/volumes" Nov 25 20:00:32 crc kubenswrapper[4759]: I1125 20:00:32.121150 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82702902-862c-4e6e-90b3-6736d46fe003" path="/var/lib/kubelet/pods/82702902-862c-4e6e-90b3-6736d46fe003/volumes" Nov 25 20:00:32 crc kubenswrapper[4759]: I1125 20:00:32.121771 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beaa5eb1-4eef-4fb3-a5ef-57fbf473804b" path="/var/lib/kubelet/pods/beaa5eb1-4eef-4fb3-a5ef-57fbf473804b/volumes" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.055686 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:35 crc kubenswrapper[4759]: E1125 20:00:35.056356 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="registry-server" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.056375 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="registry-server" Nov 25 20:00:35 crc kubenswrapper[4759]: E1125 20:00:35.056395 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="extract-content" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.056404 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="extract-content" Nov 25 20:00:35 crc kubenswrapper[4759]: E1125 20:00:35.056429 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="extract-utilities" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.056438 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="extract-utilities" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.056629 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="45828622-20a2-4839-ab4a-acbe8b3a02e6" containerName="registry-server" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.057886 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.071187 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.136394 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.136505 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n662r\" (UniqueName: \"kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.136606 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.238254 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.238345 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n662r\" (UniqueName: \"kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.238554 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.239784 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.239873 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.277250 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n662r\" (UniqueName: \"kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r\") pod \"redhat-marketplace-zvjpg\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.377947 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:35 crc kubenswrapper[4759]: I1125 20:00:35.823722 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:35 crc kubenswrapper[4759]: W1125 20:00:35.833411 4759 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67f47d0f_fe0d_40b1_9323_afb57a8cd9b7.slice/crio-2da1ca957db942e822f4769264db90e4d88c002ced579a3953b57ddf2d2f073b WatchSource:0}: Error finding container 2da1ca957db942e822f4769264db90e4d88c002ced579a3953b57ddf2d2f073b: Status 404 returned error can't find the container with id 2da1ca957db942e822f4769264db90e4d88c002ced579a3953b57ddf2d2f073b Nov 25 20:00:36 crc kubenswrapper[4759]: I1125 20:00:36.191583 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerDied","Data":"16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8"} Nov 25 20:00:36 crc kubenswrapper[4759]: I1125 20:00:36.191414 4759 generic.go:334] "Generic (PLEG): container finished" podID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerID="16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8" exitCode=0 Nov 25 20:00:36 crc kubenswrapper[4759]: I1125 20:00:36.193436 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerStarted","Data":"2da1ca957db942e822f4769264db90e4d88c002ced579a3953b57ddf2d2f073b"} Nov 25 20:00:37 crc kubenswrapper[4759]: I1125 20:00:37.203190 4759 generic.go:334] "Generic (PLEG): container finished" podID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerID="3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5" exitCode=0 Nov 25 20:00:37 crc kubenswrapper[4759]: I1125 20:00:37.203282 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerDied","Data":"3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5"} Nov 25 20:00:38 crc kubenswrapper[4759]: I1125 20:00:38.212999 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerStarted","Data":"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e"} Nov 25 20:00:38 crc kubenswrapper[4759]: I1125 20:00:38.234321 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zvjpg" podStartSLOduration=1.588586788 podStartE2EDuration="3.234298892s" podCreationTimestamp="2025-11-25 20:00:35 +0000 UTC" firstStartedPulling="2025-11-25 20:00:36.194020197 +0000 UTC m=+1626.779967898" lastFinishedPulling="2025-11-25 20:00:37.839732291 +0000 UTC m=+1628.425680002" observedRunningTime="2025-11-25 20:00:38.230988262 +0000 UTC m=+1628.816935993" watchObservedRunningTime="2025-11-25 20:00:38.234298892 +0000 UTC m=+1628.820246603" Nov 25 20:00:45 crc kubenswrapper[4759]: I1125 20:00:45.378510 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:45 crc kubenswrapper[4759]: I1125 20:00:45.379232 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:45 crc kubenswrapper[4759]: I1125 20:00:45.416258 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:46 crc kubenswrapper[4759]: I1125 20:00:46.346914 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:46 crc kubenswrapper[4759]: I1125 20:00:46.654982 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.035612 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-d4l86"] Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.041129 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-d4l86"] Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.155264 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8677c67a-c513-4719-9924-eb2a34bef984" path="/var/lib/kubelet/pods/8677c67a-c513-4719-9924-eb2a34bef984/volumes" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.294238 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zvjpg" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="registry-server" containerID="cri-o://080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e" gracePeriod=2 Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.701154 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.761132 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities\") pod \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.761308 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content\") pod \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.761386 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n662r\" (UniqueName: \"kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r\") pod \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\" (UID: \"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7\") " Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.762098 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities" (OuterVolumeSpecName: "utilities") pod "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" (UID: "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.772369 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r" (OuterVolumeSpecName: "kube-api-access-n662r") pod "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" (UID: "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7"). InnerVolumeSpecName "kube-api-access-n662r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.781599 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" (UID: "67f47d0f-fe0d-40b1-9323-afb57a8cd9b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.862995 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.863023 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:48 crc kubenswrapper[4759]: I1125 20:00:48.863037 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n662r\" (UniqueName: \"kubernetes.io/projected/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7-kube-api-access-n662r\") on node \"crc\" DevicePath \"\"" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.303364 4759 generic.go:334] "Generic (PLEG): container finished" podID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerID="080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e" exitCode=0 Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.304641 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerDied","Data":"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e"} Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.304722 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvjpg" event={"ID":"67f47d0f-fe0d-40b1-9323-afb57a8cd9b7","Type":"ContainerDied","Data":"2da1ca957db942e822f4769264db90e4d88c002ced579a3953b57ddf2d2f073b"} Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.304747 4759 scope.go:117] "RemoveContainer" containerID="080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.304660 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvjpg" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.320145 4759 scope.go:117] "RemoveContainer" containerID="3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.340510 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.343336 4759 scope.go:117] "RemoveContainer" containerID="16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.345813 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvjpg"] Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.366264 4759 scope.go:117] "RemoveContainer" containerID="080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e" Nov 25 20:00:49 crc kubenswrapper[4759]: E1125 20:00:49.366729 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e\": container with ID starting with 080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e not found: ID does not exist" containerID="080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.366760 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e"} err="failed to get container status \"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e\": rpc error: code = NotFound desc = could not find container \"080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e\": container with ID starting with 080ffacb0dd0e863d5649ae87c27266c451ef2371a2f2f40d6450165c9c5f74e not found: ID does not exist" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.366783 4759 scope.go:117] "RemoveContainer" containerID="3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5" Nov 25 20:00:49 crc kubenswrapper[4759]: E1125 20:00:49.367201 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5\": container with ID starting with 3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5 not found: ID does not exist" containerID="3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.367362 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5"} err="failed to get container status \"3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5\": rpc error: code = NotFound desc = could not find container \"3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5\": container with ID starting with 3cf49eef87937afc6cf412762e38d8cc50be44a7b535d07c075b76d3c6627cf5 not found: ID does not exist" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.367508 4759 scope.go:117] "RemoveContainer" containerID="16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8" Nov 25 20:00:49 crc kubenswrapper[4759]: E1125 20:00:49.368099 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8\": container with ID starting with 16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8 not found: ID does not exist" containerID="16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8" Nov 25 20:00:49 crc kubenswrapper[4759]: I1125 20:00:49.368241 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8"} err="failed to get container status \"16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8\": rpc error: code = NotFound desc = could not find container \"16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8\": container with ID starting with 16a23b702085960cc4d8c65998863cf90a58a1687d9dd456b99978f7342a66b8 not found: ID does not exist" Nov 25 20:00:50 crc kubenswrapper[4759]: I1125 20:00:50.118432 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" path="/var/lib/kubelet/pods/67f47d0f-fe0d-40b1-9323-afb57a8cd9b7/volumes" Nov 25 20:00:54 crc kubenswrapper[4759]: I1125 20:00:54.022328 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-g84tx"] Nov 25 20:00:54 crc kubenswrapper[4759]: I1125 20:00:54.029580 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-g84tx"] Nov 25 20:00:54 crc kubenswrapper[4759]: I1125 20:00:54.116486 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08be9db-6e35-495b-ae8e-89a8deea7695" path="/var/lib/kubelet/pods/b08be9db-6e35-495b-ae8e-89a8deea7695/volumes" Nov 25 20:00:54 crc kubenswrapper[4759]: I1125 20:00:54.895596 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:00:54 crc kubenswrapper[4759]: I1125 20:00:54.895909 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.153275 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cron-29401681-dspz7"] Nov 25 20:01:00 crc kubenswrapper[4759]: E1125 20:01:00.154110 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="extract-content" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.154125 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="extract-content" Nov 25 20:01:00 crc kubenswrapper[4759]: E1125 20:01:00.154145 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="registry-server" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.154154 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="registry-server" Nov 25 20:01:00 crc kubenswrapper[4759]: E1125 20:01:00.154171 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="extract-utilities" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.154182 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="extract-utilities" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.154398 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="67f47d0f-fe0d-40b1-9323-afb57a8cd9b7" containerName="registry-server" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.155096 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.171874 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29401681-dspz7"] Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.214673 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.215059 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2msn\" (UniqueName: \"kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.215783 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.317774 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2msn\" (UniqueName: \"kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.318122 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.318349 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.328111 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.330192 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.339527 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2msn\" (UniqueName: \"kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn\") pod \"keystone-cron-29401681-dspz7\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.476109 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:00 crc kubenswrapper[4759]: I1125 20:01:00.949885 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29401681-dspz7"] Nov 25 20:01:01 crc kubenswrapper[4759]: I1125 20:01:01.392858 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" event={"ID":"c9c821c1-82c3-404e-a219-843f034ed18e","Type":"ContainerStarted","Data":"50a12eae49ece71f1a1b4a86cf10db03a1fb64094dba03c1eaf405bdef43fc00"} Nov 25 20:01:01 crc kubenswrapper[4759]: I1125 20:01:01.394263 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" event={"ID":"c9c821c1-82c3-404e-a219-843f034ed18e","Type":"ContainerStarted","Data":"199077d63e2709a82628d5184ff2c82606cf91a87e9b650b454180d5540612f0"} Nov 25 20:01:01 crc kubenswrapper[4759]: I1125 20:01:01.426407 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" podStartSLOduration=1.42638755 podStartE2EDuration="1.42638755s" podCreationTimestamp="2025-11-25 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:01:01.425974128 +0000 UTC m=+1652.011921829" watchObservedRunningTime="2025-11-25 20:01:01.42638755 +0000 UTC m=+1652.012335251" Nov 25 20:01:03 crc kubenswrapper[4759]: I1125 20:01:03.407547 4759 generic.go:334] "Generic (PLEG): container finished" podID="c9c821c1-82c3-404e-a219-843f034ed18e" containerID="50a12eae49ece71f1a1b4a86cf10db03a1fb64094dba03c1eaf405bdef43fc00" exitCode=0 Nov 25 20:01:03 crc kubenswrapper[4759]: I1125 20:01:03.407653 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" event={"ID":"c9c821c1-82c3-404e-a219-843f034ed18e","Type":"ContainerDied","Data":"50a12eae49ece71f1a1b4a86cf10db03a1fb64094dba03c1eaf405bdef43fc00"} Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.712379 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.789382 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys\") pod \"c9c821c1-82c3-404e-a219-843f034ed18e\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.789541 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data\") pod \"c9c821c1-82c3-404e-a219-843f034ed18e\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.789617 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2msn\" (UniqueName: \"kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn\") pod \"c9c821c1-82c3-404e-a219-843f034ed18e\" (UID: \"c9c821c1-82c3-404e-a219-843f034ed18e\") " Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.801845 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn" (OuterVolumeSpecName: "kube-api-access-s2msn") pod "c9c821c1-82c3-404e-a219-843f034ed18e" (UID: "c9c821c1-82c3-404e-a219-843f034ed18e"). InnerVolumeSpecName "kube-api-access-s2msn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.801844 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c9c821c1-82c3-404e-a219-843f034ed18e" (UID: "c9c821c1-82c3-404e-a219-843f034ed18e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.829038 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data" (OuterVolumeSpecName: "config-data") pod "c9c821c1-82c3-404e-a219-843f034ed18e" (UID: "c9c821c1-82c3-404e-a219-843f034ed18e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.891707 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.891742 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2msn\" (UniqueName: \"kubernetes.io/projected/c9c821c1-82c3-404e-a219-843f034ed18e-kube-api-access-s2msn\") on node \"crc\" DevicePath \"\"" Nov 25 20:01:04 crc kubenswrapper[4759]: I1125 20:01:04.891752 4759 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c821c1-82c3-404e-a219-843f034ed18e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 20:01:05 crc kubenswrapper[4759]: I1125 20:01:05.428855 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" event={"ID":"c9c821c1-82c3-404e-a219-843f034ed18e","Type":"ContainerDied","Data":"199077d63e2709a82628d5184ff2c82606cf91a87e9b650b454180d5540612f0"} Nov 25 20:01:05 crc kubenswrapper[4759]: I1125 20:01:05.429604 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="199077d63e2709a82628d5184ff2c82606cf91a87e9b650b454180d5540612f0" Nov 25 20:01:05 crc kubenswrapper[4759]: I1125 20:01:05.429008 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29401681-dspz7" Nov 25 20:01:24 crc kubenswrapper[4759]: I1125 20:01:24.895305 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:01:24 crc kubenswrapper[4759]: I1125 20:01:24.895891 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:01:24 crc kubenswrapper[4759]: I1125 20:01:24.895937 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 20:01:24 crc kubenswrapper[4759]: I1125 20:01:24.896658 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 20:01:24 crc kubenswrapper[4759]: I1125 20:01:24.896721 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" gracePeriod=600 Nov 25 20:01:25 crc kubenswrapper[4759]: E1125 20:01:25.025207 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:01:25 crc kubenswrapper[4759]: I1125 20:01:25.595700 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" exitCode=0 Nov 25 20:01:25 crc kubenswrapper[4759]: I1125 20:01:25.595747 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad"} Nov 25 20:01:25 crc kubenswrapper[4759]: I1125 20:01:25.595814 4759 scope.go:117] "RemoveContainer" containerID="c0ee59ba56edcce7dacb963a27ad27e2f4cec0e7fab8b7d6596ce659b590adf2" Nov 25 20:01:25 crc kubenswrapper[4759]: I1125 20:01:25.596321 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:01:25 crc kubenswrapper[4759]: E1125 20:01:25.596620 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:01:31 crc kubenswrapper[4759]: I1125 20:01:31.292565 4759 scope.go:117] "RemoveContainer" containerID="833701616d0e27c7b33326731a68364129704ed5b41e662974ecf72c0611b6c0" Nov 25 20:01:31 crc kubenswrapper[4759]: I1125 20:01:31.319579 4759 scope.go:117] "RemoveContainer" containerID="7d6a59cb9456bfc3f3bc77e31251984b9c7df415b8abc3bc9cee8fb140dcb003" Nov 25 20:01:31 crc kubenswrapper[4759]: I1125 20:01:31.352398 4759 scope.go:117] "RemoveContainer" containerID="c7c2278dc4cfb0fc558be90b60957d8a2351fdf5aa9bb6bf92ee8a15ff743f1d" Nov 25 20:01:31 crc kubenswrapper[4759]: I1125 20:01:31.382805 4759 scope.go:117] "RemoveContainer" containerID="083fba8fd2dc467b2ead6b6e759a3d1895d0095fac384e46c13615a915a6063a" Nov 25 20:01:38 crc kubenswrapper[4759]: I1125 20:01:38.107321 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:01:38 crc kubenswrapper[4759]: E1125 20:01:38.107836 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:01:52 crc kubenswrapper[4759]: I1125 20:01:52.106602 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:01:52 crc kubenswrapper[4759]: E1125 20:01:52.107250 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:02:04 crc kubenswrapper[4759]: I1125 20:02:04.108215 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:02:04 crc kubenswrapper[4759]: E1125 20:02:04.109706 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:02:15 crc kubenswrapper[4759]: I1125 20:02:15.107049 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:02:15 crc kubenswrapper[4759]: E1125 20:02:15.107798 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:02:29 crc kubenswrapper[4759]: I1125 20:02:29.106878 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:02:29 crc kubenswrapper[4759]: E1125 20:02:29.108297 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:02:41 crc kubenswrapper[4759]: I1125 20:02:41.107154 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:02:41 crc kubenswrapper[4759]: E1125 20:02:41.108039 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:02:54 crc kubenswrapper[4759]: I1125 20:02:54.106795 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:02:54 crc kubenswrapper[4759]: E1125 20:02:54.107408 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.851496 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:03:05 crc kubenswrapper[4759]: E1125 20:03:05.852606 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c821c1-82c3-404e-a219-843f034ed18e" containerName="keystone-cron" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.852629 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c821c1-82c3-404e-a219-843f034ed18e" containerName="keystone-cron" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.853034 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c821c1-82c3-404e-a219-843f034ed18e" containerName="keystone-cron" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.854021 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.868539 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.969833 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.970911 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.972871 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-95wfv" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.973159 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.973199 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.978512 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 20:03:05 crc kubenswrapper[4759]: I1125 20:03:05.980372 4759 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.051702 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.051953 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.051981 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052108 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052140 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052169 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xksv\" (UniqueName: \"kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052258 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052390 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052465 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052490 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052508 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052525 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052651 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.052688 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153555 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153608 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153645 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153684 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-config\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153700 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72gj7\" (UniqueName: \"kubernetes.io/projected/379904c0-ac99-4812-af59-c043f65949bd-kube-api-access-72gj7\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153726 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153742 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153763 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/379904c0-ac99-4812-af59-c043f65949bd-openstack-config-secret\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153786 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xksv\" (UniqueName: \"kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153803 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153839 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153867 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153887 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153900 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153914 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153940 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153940 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153959 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-scripts\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.153977 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154399 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154482 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154494 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154546 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154573 4759 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154583 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154650 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154655 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154684 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.154887 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.162192 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.167084 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.172019 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.177871 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.183903 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xksv\" (UniqueName: \"kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv\") pod \"glance-default-external-api-1\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.203975 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.255068 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-scripts\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.255150 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-config\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.255165 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72gj7\" (UniqueName: \"kubernetes.io/projected/379904c0-ac99-4812-af59-c043f65949bd-kube-api-access-72gj7\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.255194 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/379904c0-ac99-4812-af59-c043f65949bd-openstack-config-secret\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.256543 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-config\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.257654 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/379904c0-ac99-4812-af59-c043f65949bd-openstack-scripts\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.260517 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/379904c0-ac99-4812-af59-c043f65949bd-openstack-config-secret\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.271246 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72gj7\" (UniqueName: \"kubernetes.io/projected/379904c0-ac99-4812-af59-c043f65949bd-kube-api-access-72gj7\") pod \"openstackclient\" (UID: \"379904c0-ac99-4812-af59-c043f65949bd\") " pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.285392 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.492757 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 20:03:06 crc kubenswrapper[4759]: I1125 20:03:06.657343 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.366430 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerStarted","Data":"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.366930 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerStarted","Data":"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.366946 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerStarted","Data":"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.366954 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerStarted","Data":"21ebafd1613c1160a32dd2622ec034d42df0e9a4dfbfc16ada3288582e0718e8"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.368155 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"379904c0-ac99-4812-af59-c043f65949bd","Type":"ContainerStarted","Data":"da364ea48d81face0f940a16d09a9655ec3a511d2e7e3029e8ed8d7305e62957"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.368178 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"379904c0-ac99-4812-af59-c043f65949bd","Type":"ContainerStarted","Data":"5c1eab213b12dd1611f4db216c50a9866c555b7fbe5bb5ee1a9eedd035996e75"} Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.400745 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.400712511 podStartE2EDuration="2.400712511s" podCreationTimestamp="2025-11-25 20:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:03:07.389485384 +0000 UTC m=+1777.975433085" watchObservedRunningTime="2025-11-25 20:03:07.400712511 +0000 UTC m=+1777.986660212" Nov 25 20:03:07 crc kubenswrapper[4759]: I1125 20:03:07.410732 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.410714876 podStartE2EDuration="2.410714876s" podCreationTimestamp="2025-11-25 20:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:03:07.404825405 +0000 UTC m=+1777.990773106" watchObservedRunningTime="2025-11-25 20:03:07.410714876 +0000 UTC m=+1777.996662577" Nov 25 20:03:09 crc kubenswrapper[4759]: I1125 20:03:09.106531 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:03:09 crc kubenswrapper[4759]: E1125 20:03:09.109154 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.204892 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.205265 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.205275 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.239631 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.244666 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.256213 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.427097 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.427413 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.427424 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.439635 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.442687 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:16 crc kubenswrapper[4759]: I1125 20:03:16.444301 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:03:24 crc kubenswrapper[4759]: I1125 20:03:24.107105 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:03:24 crc kubenswrapper[4759]: E1125 20:03:24.107836 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:03:38 crc kubenswrapper[4759]: I1125 20:03:38.106895 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:03:38 crc kubenswrapper[4759]: E1125 20:03:38.108125 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:03:52 crc kubenswrapper[4759]: I1125 20:03:52.107219 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:03:52 crc kubenswrapper[4759]: E1125 20:03:52.108551 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:05 crc kubenswrapper[4759]: I1125 20:04:05.106367 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:04:05 crc kubenswrapper[4759]: E1125 20:04:05.107378 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.322645 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.324905 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-api" containerID="cri-o://a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1" gracePeriod=30 Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.324916 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-httpd" containerID="cri-o://d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596" gracePeriod=30 Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.324775 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-log" containerID="cri-o://2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476" gracePeriod=30 Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.853420 4759 generic.go:334] "Generic (PLEG): container finished" podID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerID="d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596" exitCode=0 Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.853775 4759 generic.go:334] "Generic (PLEG): container finished" podID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerID="2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476" exitCode=143 Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.853532 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerDied","Data":"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596"} Nov 25 20:04:13 crc kubenswrapper[4759]: I1125 20:04:13.853827 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerDied","Data":"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476"} Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.205435 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311686 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311758 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311792 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311842 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311858 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311893 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311917 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xksv\" (UniqueName: \"kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311946 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311962 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311973 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.311990 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312015 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312036 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312064 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data\") pod \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\" (UID: \"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7\") " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312518 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys" (OuterVolumeSpecName: "sys") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312518 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev" (OuterVolumeSpecName: "dev") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312535 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312564 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312699 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312701 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run" (OuterVolumeSpecName: "run") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312736 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.312786 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.313198 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs" (OuterVolumeSpecName: "logs") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.317561 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.317852 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts" (OuterVolumeSpecName: "scripts") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.317856 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.319909 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv" (OuterVolumeSpecName: "kube-api-access-8xksv") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "kube-api-access-8xksv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.387945 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data" (OuterVolumeSpecName: "config-data") pod "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" (UID: "1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414372 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414415 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414424 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414432 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414455 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414568 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414614 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414628 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414640 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414679 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414692 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xksv\" (UniqueName: \"kubernetes.io/projected/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-kube-api-access-8xksv\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414707 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-dev\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414718 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-sys\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.414730 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.429350 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.432105 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.516728 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.517204 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.660699 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wxlrw"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.667111 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wxlrw"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.751841 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance6696-account-delete-xw9p8"] Nov 25 20:04:14 crc kubenswrapper[4759]: E1125 20:04:14.752218 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-api" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752242 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-api" Nov 25 20:04:14 crc kubenswrapper[4759]: E1125 20:04:14.752262 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-httpd" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752270 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-httpd" Nov 25 20:04:14 crc kubenswrapper[4759]: E1125 20:04:14.752278 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-log" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752286 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-log" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752479 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-api" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752496 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-httpd" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.752519 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerName="glance-log" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.753090 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.759079 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6696-account-delete-xw9p8"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.875236 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.875681 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-log" containerID="cri-o://e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.876163 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-httpd" containerID="cri-o://be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.876142 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-api" containerID="cri-o://7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.883422 4759 generic.go:334] "Generic (PLEG): container finished" podID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" containerID="a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1" exitCode=0 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.883513 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerDied","Data":"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1"} Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.883547 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.883568 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7","Type":"ContainerDied","Data":"21ebafd1613c1160a32dd2622ec034d42df0e9a4dfbfc16ada3288582e0718e8"} Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.883591 4759 scope.go:117] "RemoveContainer" containerID="a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.900152 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.912497 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940168rjt67"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.922851 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.922916 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlxjb\" (UniqueName: \"kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.932341 4759 scope.go:117] "RemoveContainer" containerID="d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.942206 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.958714 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.964458 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.964717 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-log" containerID="cri-o://6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.965890 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-api" containerID="cri-o://583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.965997 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-httpd" containerID="cri-o://11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" gracePeriod=30 Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.979722 4759 scope.go:117] "RemoveContainer" containerID="2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476" Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.981703 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr"] Nov 25 20:04:14 crc kubenswrapper[4759]: I1125 20:04:14.990843 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940168dsllr"] Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.024897 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlxjb\" (UniqueName: \"kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.025003 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.025697 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048088 4759 scope.go:117] "RemoveContainer" containerID="a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1" Nov 25 20:04:15 crc kubenswrapper[4759]: E1125 20:04:15.048418 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1\": container with ID starting with a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1 not found: ID does not exist" containerID="a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048467 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1"} err="failed to get container status \"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1\": rpc error: code = NotFound desc = could not find container \"a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1\": container with ID starting with a563409f8ae877227f74d857c9f77d53629e63d796c494a5d9254fe7a8f4d0a1 not found: ID does not exist" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048494 4759 scope.go:117] "RemoveContainer" containerID="d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048623 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlxjb\" (UniqueName: \"kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb\") pod \"glance6696-account-delete-xw9p8\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:15 crc kubenswrapper[4759]: E1125 20:04:15.048739 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596\": container with ID starting with d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596 not found: ID does not exist" containerID="d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048758 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596"} err="failed to get container status \"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596\": rpc error: code = NotFound desc = could not find container \"d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596\": container with ID starting with d7edf8c526882b3f8b201df6815ac1cda08226e4395759e8b5448193ffce0596 not found: ID does not exist" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048772 4759 scope.go:117] "RemoveContainer" containerID="2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476" Nov 25 20:04:15 crc kubenswrapper[4759]: E1125 20:04:15.048977 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476\": container with ID starting with 2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476 not found: ID does not exist" containerID="2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.048994 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476"} err="failed to get container status \"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476\": rpc error: code = NotFound desc = could not find container \"2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476\": container with ID starting with 2d0616a9e75368ba83f153f3af58aa86b436f428c4673f5532f00594881d4476 not found: ID does not exist" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.119417 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.563286 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6696-account-delete-xw9p8"] Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.761671 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.860056 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894775 4759 generic.go:334] "Generic (PLEG): container finished" podID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" exitCode=0 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894808 4759 generic.go:334] "Generic (PLEG): container finished" podID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" exitCode=0 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894815 4759 generic.go:334] "Generic (PLEG): container finished" podID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" exitCode=143 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894835 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894877 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerDied","Data":"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894958 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerDied","Data":"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.894974 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerDied","Data":"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.895022 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"11fcc57f-260e-4cdd-a2a5-acfce9c9b676","Type":"ContainerDied","Data":"33e47a7835419fd7c0f3128c861d7ddd973c352de46ad21931b88084a597183e"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.895054 4759 scope.go:117] "RemoveContainer" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897145 4759 generic.go:334] "Generic (PLEG): container finished" podID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" exitCode=0 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897168 4759 generic.go:334] "Generic (PLEG): container finished" podID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" exitCode=0 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897175 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897190 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerDied","Data":"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897219 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerDied","Data":"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897230 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerDied","Data":"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897175 4759 generic.go:334] "Generic (PLEG): container finished" podID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" exitCode=143 Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.897345 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"29d07b8f-c11f-4180-a5f5-8e0735e59389","Type":"ContainerDied","Data":"cc2e389b12415c48e3c84f6618716442a89d3b83a6b6b9608550a18e0c947c1e"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.898962 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" event={"ID":"7655c25c-b868-45fc-a8f9-b252f93f12ae","Type":"ContainerStarted","Data":"cec487aa8556481a16785014f489b6e2fc841d6d78aa94f67a063235963404be"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.899003 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" event={"ID":"7655c25c-b868-45fc-a8f9-b252f93f12ae","Type":"ContainerStarted","Data":"6b232bb42e4bf396bf88c7402af7fbdd5abc9c47e0cd0d69229cd2b1ca19014f"} Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.912260 4759 scope.go:117] "RemoveContainer" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.917618 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" podStartSLOduration=1.917600599 podStartE2EDuration="1.917600599s" podCreationTimestamp="2025-11-25 20:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 20:04:15.914269668 +0000 UTC m=+1846.500217369" watchObservedRunningTime="2025-11-25 20:04:15.917600599 +0000 UTC m=+1846.503548300" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.935207 4759 scope.go:117] "RemoveContainer" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940454 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940487 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940506 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940521 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940537 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940554 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940572 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940588 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940604 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940623 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940640 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940653 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940669 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940681 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940709 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940729 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbwhd\" (UniqueName: \"kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940749 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940774 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940799 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940816 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nvtv\" (UniqueName: \"kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940839 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940858 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940873 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940890 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940913 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940929 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts\") pod \"29d07b8f-c11f-4180-a5f5-8e0735e59389\" (UID: \"29d07b8f-c11f-4180-a5f5-8e0735e59389\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940959 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.940972 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev\") pod \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\" (UID: \"11fcc57f-260e-4cdd-a2a5-acfce9c9b676\") " Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.941678 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942769 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942815 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942770 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys" (OuterVolumeSpecName: "sys") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942844 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run" (OuterVolumeSpecName: "run") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942862 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.942888 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943234 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs" (OuterVolumeSpecName: "logs") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943516 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run" (OuterVolumeSpecName: "run") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943543 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943475 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943786 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev" (OuterVolumeSpecName: "dev") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943898 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs" (OuterVolumeSpecName: "logs") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.943984 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.944034 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys" (OuterVolumeSpecName: "sys") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.944084 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev" (OuterVolumeSpecName: "dev") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.944113 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.944142 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.946974 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.948376 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.948439 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd" (OuterVolumeSpecName: "kube-api-access-nbwhd") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "kube-api-access-nbwhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.948473 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.948779 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.949606 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts" (OuterVolumeSpecName: "scripts") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.950017 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv" (OuterVolumeSpecName: "kube-api-access-4nvtv") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "kube-api-access-4nvtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:04:15 crc kubenswrapper[4759]: I1125 20:04:15.950628 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts" (OuterVolumeSpecName: "scripts") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.000887 4759 scope.go:117] "RemoveContainer" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.001907 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": container with ID starting with 7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947 not found: ID does not exist" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.001973 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947"} err="failed to get container status \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": rpc error: code = NotFound desc = could not find container \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": container with ID starting with 7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002007 4759 scope.go:117] "RemoveContainer" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.002288 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": container with ID starting with be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96 not found: ID does not exist" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002318 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96"} err="failed to get container status \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": rpc error: code = NotFound desc = could not find container \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": container with ID starting with be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002339 4759 scope.go:117] "RemoveContainer" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.002685 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": container with ID starting with e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591 not found: ID does not exist" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002705 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591"} err="failed to get container status \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": rpc error: code = NotFound desc = could not find container \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": container with ID starting with e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002716 4759 scope.go:117] "RemoveContainer" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002957 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947"} err="failed to get container status \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": rpc error: code = NotFound desc = could not find container \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": container with ID starting with 7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.002999 4759 scope.go:117] "RemoveContainer" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.003277 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96"} err="failed to get container status \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": rpc error: code = NotFound desc = could not find container \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": container with ID starting with be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.003317 4759 scope.go:117] "RemoveContainer" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006483 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591"} err="failed to get container status \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": rpc error: code = NotFound desc = could not find container \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": container with ID starting with e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006508 4759 scope.go:117] "RemoveContainer" containerID="7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006735 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947"} err="failed to get container status \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": rpc error: code = NotFound desc = could not find container \"7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947\": container with ID starting with 7314be5f2e57020c6198bb86fd12ff7d303c8995d4af5a7eee808c6b2d025947 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006752 4759 scope.go:117] "RemoveContainer" containerID="be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006914 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96"} err="failed to get container status \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": rpc error: code = NotFound desc = could not find container \"be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96\": container with ID starting with be61e24ffbe6d88c61d9ea97c27de0710648011d5f24515ffc1702f1c83d2e96 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.006930 4759 scope.go:117] "RemoveContainer" containerID="e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.007076 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591"} err="failed to get container status \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": rpc error: code = NotFound desc = could not find container \"e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591\": container with ID starting with e8e0b0748db8b7fe4c9a15355e79c01ba2c7ffd87e72e7d90fa88953a38db591 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.007093 4759 scope.go:117] "RemoveContainer" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.012285 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data" (OuterVolumeSpecName: "config-data") pod "11fcc57f-260e-4cdd-a2a5-acfce9c9b676" (UID: "11fcc57f-260e-4cdd-a2a5-acfce9c9b676"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.020516 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data" (OuterVolumeSpecName: "config-data") pod "29d07b8f-c11f-4180-a5f5-8e0735e59389" (UID: "29d07b8f-c11f-4180-a5f5-8e0735e59389"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.026775 4759 scope.go:117] "RemoveContainer" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041833 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-sys\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041865 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041893 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041905 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nvtv\" (UniqueName: \"kubernetes.io/projected/29d07b8f-c11f-4180-a5f5-8e0735e59389-kube-api-access-4nvtv\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041915 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041923 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041931 4759 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041940 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041947 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041955 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041961 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-logs\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041969 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-dev\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041976 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041988 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.041996 4759 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-sys\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042008 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042016 4759 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042023 4759 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042032 4759 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-dev\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042043 4759 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042052 4759 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042059 4759 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29d07b8f-c11f-4180-a5f5-8e0735e59389-logs\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042067 4759 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042075 4759 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042086 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042096 4759 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29d07b8f-c11f-4180-a5f5-8e0735e59389-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042104 4759 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/29d07b8f-c11f-4180-a5f5-8e0735e59389-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.042112 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbwhd\" (UniqueName: \"kubernetes.io/projected/11fcc57f-260e-4cdd-a2a5-acfce9c9b676-kube-api-access-nbwhd\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.064350 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.064839 4759 scope.go:117] "RemoveContainer" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.067892 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.075092 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.076319 4759 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.079055 4759 scope.go:117] "RemoveContainer" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.079411 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": container with ID starting with 583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c not found: ID does not exist" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.079474 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c"} err="failed to get container status \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": rpc error: code = NotFound desc = could not find container \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": container with ID starting with 583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.079528 4759 scope.go:117] "RemoveContainer" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.079802 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": container with ID starting with 11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521 not found: ID does not exist" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.079823 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521"} err="failed to get container status \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": rpc error: code = NotFound desc = could not find container \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": container with ID starting with 11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.079857 4759 scope.go:117] "RemoveContainer" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.080206 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": container with ID starting with 6d951225d64857a070798003891e8861841876830260619f0a4a855257279607 not found: ID does not exist" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.080304 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607"} err="failed to get container status \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": rpc error: code = NotFound desc = could not find container \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": container with ID starting with 6d951225d64857a070798003891e8861841876830260619f0a4a855257279607 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.080376 4759 scope.go:117] "RemoveContainer" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.080687 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c"} err="failed to get container status \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": rpc error: code = NotFound desc = could not find container \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": container with ID starting with 583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.080710 4759 scope.go:117] "RemoveContainer" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081069 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521"} err="failed to get container status \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": rpc error: code = NotFound desc = could not find container \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": container with ID starting with 11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081164 4759 scope.go:117] "RemoveContainer" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081433 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607"} err="failed to get container status \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": rpc error: code = NotFound desc = could not find container \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": container with ID starting with 6d951225d64857a070798003891e8861841876830260619f0a4a855257279607 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081521 4759 scope.go:117] "RemoveContainer" containerID="583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081776 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c"} err="failed to get container status \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": rpc error: code = NotFound desc = could not find container \"583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c\": container with ID starting with 583c95d2f741e9de1951f6388757d521ba1ce2964b919afc4b67a28472c27d3c not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.081861 4759 scope.go:117] "RemoveContainer" containerID="11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.082095 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521"} err="failed to get container status \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": rpc error: code = NotFound desc = could not find container \"11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521\": container with ID starting with 11bf4c690f57b086f5ce3dcb4062e24702badbaee7712391217de3c4ea427521 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.082168 4759 scope.go:117] "RemoveContainer" containerID="6d951225d64857a070798003891e8861841876830260619f0a4a855257279607" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.082508 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607"} err="failed to get container status \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": rpc error: code = NotFound desc = could not find container \"6d951225d64857a070798003891e8861841876830260619f0a4a855257279607\": container with ID starting with 6d951225d64857a070798003891e8861841876830260619f0a4a855257279607 not found: ID does not exist" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.108660 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:04:16 crc kubenswrapper[4759]: E1125 20:04:16.108929 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.115915 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142" path="/var/lib/kubelet/pods/0abf4b9b-8f46-4cd7-8c13-f0c6dc1cd142/volumes" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.116789 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7" path="/var/lib/kubelet/pods/1c6eccc2-e8a2-4eb9-ab61-a1fd046bbaa7/volumes" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.117574 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e78b547-f232-41e3-a1b3-df5415fec073" path="/var/lib/kubelet/pods/8e78b547-f232-41e3-a1b3-df5415fec073/volumes" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.118861 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61252ae-a6de-47a2-9355-d9392583883c" path="/var/lib/kubelet/pods/c61252ae-a6de-47a2-9355-d9392583883c/volumes" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.143112 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.143142 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.143154 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.143166 4759 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.216002 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.224333 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.241771 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.249519 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.911477 4759 generic.go:334] "Generic (PLEG): container finished" podID="7655c25c-b868-45fc-a8f9-b252f93f12ae" containerID="cec487aa8556481a16785014f489b6e2fc841d6d78aa94f67a063235963404be" exitCode=0 Nov 25 20:04:16 crc kubenswrapper[4759]: I1125 20:04:16.911515 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" event={"ID":"7655c25c-b868-45fc-a8f9-b252f93f12ae","Type":"ContainerDied","Data":"cec487aa8556481a16785014f489b6e2fc841d6d78aa94f67a063235963404be"} Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.123701 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" path="/var/lib/kubelet/pods/11fcc57f-260e-4cdd-a2a5-acfce9c9b676/volumes" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.125298 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" path="/var/lib/kubelet/pods/29d07b8f-c11f-4180-a5f5-8e0735e59389/volumes" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.306648 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.474787 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts\") pod \"7655c25c-b868-45fc-a8f9-b252f93f12ae\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.474852 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlxjb\" (UniqueName: \"kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb\") pod \"7655c25c-b868-45fc-a8f9-b252f93f12ae\" (UID: \"7655c25c-b868-45fc-a8f9-b252f93f12ae\") " Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.475437 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7655c25c-b868-45fc-a8f9-b252f93f12ae" (UID: "7655c25c-b868-45fc-a8f9-b252f93f12ae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.479903 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb" (OuterVolumeSpecName: "kube-api-access-nlxjb") pod "7655c25c-b868-45fc-a8f9-b252f93f12ae" (UID: "7655c25c-b868-45fc-a8f9-b252f93f12ae"). InnerVolumeSpecName "kube-api-access-nlxjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.576501 4759 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7655c25c-b868-45fc-a8f9-b252f93f12ae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.576538 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlxjb\" (UniqueName: \"kubernetes.io/projected/7655c25c-b868-45fc-a8f9-b252f93f12ae-kube-api-access-nlxjb\") on node \"crc\" DevicePath \"\"" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.933116 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" event={"ID":"7655c25c-b868-45fc-a8f9-b252f93f12ae","Type":"ContainerDied","Data":"6b232bb42e4bf396bf88c7402af7fbdd5abc9c47e0cd0d69229cd2b1ca19014f"} Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.933626 4759 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b232bb42e4bf396bf88c7402af7fbdd5abc9c47e0cd0d69229cd2b1ca19014f" Nov 25 20:04:18 crc kubenswrapper[4759]: I1125 20:04:18.933208 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6696-account-delete-xw9p8" Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.779008 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-lvtvr"] Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.787360 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-lvtvr"] Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.802775 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance6696-account-delete-xw9p8"] Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.810812 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-6696-account-create-update-pxz2s"] Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.818034 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance6696-account-delete-xw9p8"] Nov 25 20:04:19 crc kubenswrapper[4759]: I1125 20:04:19.824522 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-6696-account-create-update-pxz2s"] Nov 25 20:04:20 crc kubenswrapper[4759]: I1125 20:04:20.117168 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7655c25c-b868-45fc-a8f9-b252f93f12ae" path="/var/lib/kubelet/pods/7655c25c-b868-45fc-a8f9-b252f93f12ae/volumes" Nov 25 20:04:20 crc kubenswrapper[4759]: I1125 20:04:20.117933 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf" path="/var/lib/kubelet/pods/7f410cd7-ce9c-4df1-a0b6-32b54fe82bdf/volumes" Nov 25 20:04:20 crc kubenswrapper[4759]: I1125 20:04:20.118685 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="965de3f2-28e4-4190-8528-305cd8978d33" path="/var/lib/kubelet/pods/965de3f2-28e4-4190-8528-305cd8978d33/volumes" Nov 25 20:04:27 crc kubenswrapper[4759]: I1125 20:04:27.107487 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:04:27 crc kubenswrapper[4759]: E1125 20:04:27.108364 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:31 crc kubenswrapper[4759]: I1125 20:04:31.495200 4759 scope.go:117] "RemoveContainer" containerID="372db637c6707ae6010d0dca2eca2783cd94a1989227a8f88c8a1ab20f8140b5" Nov 25 20:04:31 crc kubenswrapper[4759]: I1125 20:04:31.521323 4759 scope.go:117] "RemoveContainer" containerID="837ffc79cf7e9d10c681676a19def92de843ae76b9d608076740c1bc3beab99b" Nov 25 20:04:31 crc kubenswrapper[4759]: I1125 20:04:31.606650 4759 scope.go:117] "RemoveContainer" containerID="909590ff9c2c5257bc0400f06753c5963581c0cecf55110a06839ee6dd296abf" Nov 25 20:04:39 crc kubenswrapper[4759]: I1125 20:04:39.106751 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:04:39 crc kubenswrapper[4759]: E1125 20:04:39.107604 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:50 crc kubenswrapper[4759]: I1125 20:04:50.126345 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:04:50 crc kubenswrapper[4759]: E1125 20:04:50.127393 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.384877 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s6tfq/must-gather-5hrqv"] Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.386892 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.387035 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.387157 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.387252 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.387353 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.387478 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.387617 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.387726 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.387863 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.388006 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.388262 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7655c25c-b868-45fc-a8f9-b252f93f12ae" containerName="mariadb-account-delete" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.388373 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="7655c25c-b868-45fc-a8f9-b252f93f12ae" containerName="mariadb-account-delete" Nov 25 20:04:53 crc kubenswrapper[4759]: E1125 20:04:53.388588 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.388711 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389087 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389215 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389362 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="7655c25c-b868-45fc-a8f9-b252f93f12ae" containerName="mariadb-account-delete" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389524 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-log" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389650 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="29d07b8f-c11f-4180-a5f5-8e0735e59389" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389760 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-httpd" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.389874 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fcc57f-260e-4cdd-a2a5-acfce9c9b676" containerName="glance-api" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.391184 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.393907 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s6tfq"/"kube-root-ca.crt" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.394133 4759 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s6tfq"/"openshift-service-ca.crt" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.406687 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s6tfq/must-gather-5hrqv"] Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.518235 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.518335 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjgb8\" (UniqueName: \"kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.620034 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.620517 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjgb8\" (UniqueName: \"kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.620538 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.644348 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjgb8\" (UniqueName: \"kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8\") pod \"must-gather-5hrqv\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:53 crc kubenswrapper[4759]: I1125 20:04:53.712806 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:04:54 crc kubenswrapper[4759]: I1125 20:04:54.178869 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s6tfq/must-gather-5hrqv"] Nov 25 20:04:54 crc kubenswrapper[4759]: I1125 20:04:54.195417 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 20:04:54 crc kubenswrapper[4759]: I1125 20:04:54.229074 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" event={"ID":"b0772673-9511-4e08-85ca-853958cf8519","Type":"ContainerStarted","Data":"233fc4a638a964f6bd9642acd41191c8911d71683b62bdccc8cef612cbc749a0"} Nov 25 20:04:58 crc kubenswrapper[4759]: I1125 20:04:58.258488 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" event={"ID":"b0772673-9511-4e08-85ca-853958cf8519","Type":"ContainerStarted","Data":"4ce5b86703d3c675bd30827ca984fd55b609fe4d406b649303c9ebc7bb7b164e"} Nov 25 20:04:59 crc kubenswrapper[4759]: I1125 20:04:59.266913 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" event={"ID":"b0772673-9511-4e08-85ca-853958cf8519","Type":"ContainerStarted","Data":"3ed8d47158d6e1a28ec3fbfe1f6b9ec84799fc9c3a90e61c97871ca2ce2f97da"} Nov 25 20:04:59 crc kubenswrapper[4759]: I1125 20:04:59.284703 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" podStartSLOduration=2.610519589 podStartE2EDuration="6.284683578s" podCreationTimestamp="2025-11-25 20:04:53 +0000 UTC" firstStartedPulling="2025-11-25 20:04:54.19531166 +0000 UTC m=+1884.781259381" lastFinishedPulling="2025-11-25 20:04:57.869475669 +0000 UTC m=+1888.455423370" observedRunningTime="2025-11-25 20:04:59.280380393 +0000 UTC m=+1889.866328094" watchObservedRunningTime="2025-11-25 20:04:59.284683578 +0000 UTC m=+1889.870631279" Nov 25 20:05:04 crc kubenswrapper[4759]: I1125 20:05:04.106588 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:05:04 crc kubenswrapper[4759]: E1125 20:05:04.107278 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:05:15 crc kubenswrapper[4759]: I1125 20:05:15.107144 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:05:15 crc kubenswrapper[4759]: E1125 20:05:15.107859 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:05:27 crc kubenswrapper[4759]: I1125 20:05:27.106606 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:05:27 crc kubenswrapper[4759]: E1125 20:05:27.107367 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:05:30 crc kubenswrapper[4759]: I1125 20:05:30.717127 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/util/0.log" Nov 25 20:05:30 crc kubenswrapper[4759]: I1125 20:05:30.835223 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/util/0.log" Nov 25 20:05:30 crc kubenswrapper[4759]: I1125 20:05:30.857987 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/pull/0.log" Nov 25 20:05:30 crc kubenswrapper[4759]: I1125 20:05:30.874059 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.010629 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/extract/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.051601 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/util/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.053081 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fk2w7w_11846abc-749c-4559-8f1b-38f0df47237b/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.177730 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/util/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.325103 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/util/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.343009 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.377740 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.515610 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/util/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.539324 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/extract/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.558315 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3cdca3774e130fb88daeefc86062465897fe8a1e81060e134279a7da68vdtjl_b4fc01b1-0131-4fa2-84bf-a4b000301ea8/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.665674 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/util/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.846202 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.876982 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/pull/0.log" Nov 25 20:05:31 crc kubenswrapper[4759]: I1125 20:05:31.881711 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.003642 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.018575 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/pull/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.071781 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dtqbvp_b11f3e19-ffc6-4a83-83b1-25467cbcbc4c/extract/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.178108 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.336731 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.352784 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/pull/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.364814 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/pull/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.486902 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.538152 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/pull/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.554108 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bkfwh8_b160ce8c-b3e2-4934-8f0c-9a1ff3926b0f/extract/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.667882 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.815661 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/pull/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.821377 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/util/0.log" Nov 25 20:05:32 crc kubenswrapper[4759]: I1125 20:05:32.896128 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.119015 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.147927 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/extract/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.148882 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907628t_ca467d28-bd2c-46a1-a27a-e34145ec978f/util/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.294244 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/util/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.442016 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.451834 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.473937 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/util/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.627998 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.661646 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/util/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.707632 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368chw72n_25813c0d-014c-48be-8468-751289d63d85/extract/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.716981 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/util/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.878871 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.893394 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/pull/0.log" Nov 25 20:05:33 crc kubenswrapper[4759]: I1125 20:05:33.899873 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/util/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.028682 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/pull/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.038045 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/util/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.101853 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3ffcrv_a22e732e-34e5-4d03-b491-72522fb7f9e4/extract/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.158729 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-85b98645d7-9h75s_3843646a-ca2b-420e-b03b-a41a0e8e755a/manager/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.197947 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-cg7g2_c0e21472-d327-4f63-a8a5-be294881a356/registry-server/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.297220 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7444b5967c-rfslf_47df9f04-b8f4-4a25-b578-3d7fe9cb402e/manager/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.374812 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-7647x_85df7f59-3c49-4ece-a44e-c6641ee5da2f/registry-server/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.469314 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-59d5f8f59b-7hjcc_979866bf-3d26-40f9-abfb-81aada52864c/kube-rbac-proxy/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.539330 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-59d5f8f59b-7hjcc_979866bf-3d26-40f9-abfb-81aada52864c/manager/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.544902 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-5mwnr_224d5f03-e4d9-4652-bfc0-e7eb6bb452c1/registry-server/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.696666 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f964ccf8d-kwvtg_19f1aaf0-6525-4ea2-96fa-8dc16e3e2871/manager/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.738493 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-jhrcq_02b44d5e-17d3-41e5-91fb-c5bae66cf31c/registry-server/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.798490 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-847fcc49c9-t9qj8_4afc0c17-9efa-4f7c-bd72-c2a3091eeddd/manager/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.936522 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-whf2r_d3320894-675c-4b39-a183-251eda3f4cce/registry-server/0.log" Nov 25 20:05:34 crc kubenswrapper[4759]: I1125 20:05:34.986052 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-26bxr_3d48201b-dacc-49bd-9d62-c7dbb87af2fa/operator/0.log" Nov 25 20:05:35 crc kubenswrapper[4759]: I1125 20:05:35.054547 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-pzj5w_59198c27-1433-45fd-b774-7e95d852ee68/registry-server/0.log" Nov 25 20:05:35 crc kubenswrapper[4759]: I1125 20:05:35.120576 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9975d566b-gc7jj_ec15c425-7822-499c-8bea-f9bad0f431a8/manager/0.log" Nov 25 20:05:35 crc kubenswrapper[4759]: I1125 20:05:35.204175 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-2v6hm_94edb074-48ac-4b2f-ad03-6fe590c96b5c/registry-server/0.log" Nov 25 20:05:39 crc kubenswrapper[4759]: I1125 20:05:39.107219 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:05:39 crc kubenswrapper[4759]: E1125 20:05:39.107675 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:05:48 crc kubenswrapper[4759]: I1125 20:05:48.989328 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7mlsn_32fdca48-e90c-4f5e-8524-3abaeb2f1bfb/control-plane-machine-set-operator/0.log" Nov 25 20:05:49 crc kubenswrapper[4759]: I1125 20:05:49.145826 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mqv76_0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16/kube-rbac-proxy/0.log" Nov 25 20:05:49 crc kubenswrapper[4759]: I1125 20:05:49.167728 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mqv76_0e2d399e-1cf7-4b82-bdd5-d5c1f59a2b16/machine-api-operator/0.log" Nov 25 20:05:51 crc kubenswrapper[4759]: I1125 20:05:51.107018 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:05:51 crc kubenswrapper[4759]: E1125 20:05:51.107538 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:06:03 crc kubenswrapper[4759]: I1125 20:06:03.805981 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-76794_bc126569-65d5-4b4e-80cd-5b073b4faca9/kube-rbac-proxy/0.log" Nov 25 20:06:03 crc kubenswrapper[4759]: I1125 20:06:03.888350 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-76794_bc126569-65d5-4b4e-80cd-5b073b4faca9/controller/0.log" Nov 25 20:06:03 crc kubenswrapper[4759]: I1125 20:06:03.951067 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-frr-files/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.176143 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-reloader/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.181721 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-reloader/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.185160 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-frr-files/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.211626 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-metrics/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.369694 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-frr-files/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.384960 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-reloader/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.390711 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-metrics/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.408756 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-metrics/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.559278 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-metrics/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.566455 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-frr-files/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.572981 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/controller/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.634224 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/cp-reloader/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.750815 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/frr-metrics/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.781012 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/kube-rbac-proxy/0.log" Nov 25 20:06:04 crc kubenswrapper[4759]: I1125 20:06:04.825632 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/kube-rbac-proxy-frr/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.039657 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/reloader/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.046069 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-989kk_95d5dc45-33bb-4278-a426-7534af845b52/frr-k8s-webhook-server/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.106289 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:06:05 crc kubenswrapper[4759]: E1125 20:06:05.106527 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.175262 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8tdxq_f9bf8e8a-6a2f-4f2d-aedf-98b57a83974a/frr/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.226506 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-747d477645-tgk7n_29fe3fc7-edb3-4ffd-b213-b53a2a055a1b/manager/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.451820 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84b47d7745-jnp5n_94d17030-8068-486d-906f-558adb93cdfb/webhook-server/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.466856 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bcmbj_d395fdc1-d2ad-413a-8cf9-8e6cf31e324e/kube-rbac-proxy/0.log" Nov 25 20:06:05 crc kubenswrapper[4759]: I1125 20:06:05.616891 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bcmbj_d395fdc1-d2ad-413a-8cf9-8e6cf31e324e/speaker/0.log" Nov 25 20:06:16 crc kubenswrapper[4759]: I1125 20:06:16.106666 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:06:16 crc kubenswrapper[4759]: E1125 20:06:16.107406 4759 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfsr5_openshift-machine-config-operator(07c1b558-ca9e-4edc-9420-05750ceee3a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.245121 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-cron-29401681-dspz7_c9c821c1-82c3-404e-a219-843f034ed18e/keystone-cron/0.log" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.344651 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-5b597bd596-ghgvg_b9a9088f-12be-4169-8bd6-2f641703df0f/keystone-api/0.log" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.591956 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6deae4a9-9a28-48fa-800d-9749ac2dcf09/mysql-bootstrap/0.log" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.696821 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6deae4a9-9a28-48fa-800d-9749ac2dcf09/galera/0.log" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.718880 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6deae4a9-9a28-48fa-800d-9749ac2dcf09/mysql-bootstrap/0.log" Nov 25 20:06:18 crc kubenswrapper[4759]: I1125 20:06:18.867599 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_5e10fc44-a367-4da8-9414-9aa286b2d9c1/mysql-bootstrap/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.101904 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_5e10fc44-a367-4da8-9414-9aa286b2d9c1/galera/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.171189 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_5e10fc44-a367-4da8-9414-9aa286b2d9c1/mysql-bootstrap/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.313238 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3c9d592c-a1e4-4e93-968d-f98c10049b84/mysql-bootstrap/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.519110 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3c9d592c-a1e4-4e93-968d-f98c10049b84/mysql-bootstrap/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.521086 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3c9d592c-a1e4-4e93-968d-f98c10049b84/galera/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.794883 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_379904c0-ac99-4812-af59-c043f65949bd/openstackclient/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.801209 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_fb645c67-9f04-4bf4-a350-00cea6872805/memcached/0.log" Nov 25 20:06:19 crc kubenswrapper[4759]: I1125 20:06:19.875055 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_de35a78c-a4e0-4610-89a3-0802b6bb3dfd/setup-container/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.024853 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_de35a78c-a4e0-4610-89a3-0802b6bb3dfd/setup-container/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.104959 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-zlfnc_f4b24bd8-f201-4cbb-8a6c-1821e1c1002c/proxy-httpd/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.118569 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_de35a78c-a4e0-4610-89a3-0802b6bb3dfd/rabbitmq/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.202312 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-zlfnc_f4b24bd8-f201-4cbb-8a6c-1821e1c1002c/proxy-server/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.274829 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-vgfs7_d763ac10-2547-4a41-a788-c784d9a0bd84/swift-ring-rebalance/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.370243 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/account-auditor/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.432646 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/account-replicator/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.439055 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/account-reaper/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.499303 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/account-server/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.548632 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/container-auditor/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.626286 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/container-server/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.631009 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/container-replicator/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.708490 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/container-updater/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.737239 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/object-auditor/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.778778 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/object-expirer/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.825530 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/object-replicator/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.887045 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/object-server/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.947609 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/object-updater/0.log" Nov 25 20:06:20 crc kubenswrapper[4759]: I1125 20:06:20.958499 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/rsync/0.log" Nov 25 20:06:21 crc kubenswrapper[4759]: I1125 20:06:21.010009 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_07cbf89a-8a0b-48ab-9e23-52226f362fbc/swift-recon-cron/0.log" Nov 25 20:06:31 crc kubenswrapper[4759]: I1125 20:06:31.106925 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:06:31 crc kubenswrapper[4759]: I1125 20:06:31.717820 4759 scope.go:117] "RemoveContainer" containerID="9b5c8ec545cdd11486c9d79b81e0f8e3ed9dca45e52b6179129e140f3fd8cea1" Nov 25 20:06:31 crc kubenswrapper[4759]: I1125 20:06:31.752894 4759 scope.go:117] "RemoveContainer" containerID="dd563ef69d90ec4d702b8650a3d976c0dd6b6061ee8ac82425ad4bab117a1ffa" Nov 25 20:06:31 crc kubenswrapper[4759]: I1125 20:06:31.909860 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"34527af12c7198aeb773451b05c7977f1f065c3b0959196146823ccb0c85db94"} Nov 25 20:06:32 crc kubenswrapper[4759]: I1125 20:06:32.736502 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-utilities/0.log" Nov 25 20:06:32 crc kubenswrapper[4759]: I1125 20:06:32.913716 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-utilities/0.log" Nov 25 20:06:32 crc kubenswrapper[4759]: I1125 20:06:32.919922 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-content/0.log" Nov 25 20:06:32 crc kubenswrapper[4759]: I1125 20:06:32.925395 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-content/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.072171 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-content/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.086199 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/extract-utilities/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.283775 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-utilities/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.410328 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ccnmk_56bef130-bc8b-49c5-b9e1-c95867515bbc/registry-server/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.470506 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-content/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.531956 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-content/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.532042 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-utilities/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.662662 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-utilities/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.719782 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/extract-content/0.log" Nov 25 20:06:33 crc kubenswrapper[4759]: I1125 20:06:33.836340 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/util/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.067566 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5scxm_00a8d174-0d8b-4841-ace2-958422c2d299/registry-server/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.080027 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/pull/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.114820 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/util/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.115332 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/pull/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.312869 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/util/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.319643 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/extract/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.350152 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ljwnq_a4147e62-c53d-46d2-89f2-bc1ccb3859e7/pull/0.log" Nov 25 20:06:34 crc kubenswrapper[4759]: I1125 20:06:34.488190 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tvgvw_95673284-6f81-4c46-8fa6-825dc24d77b1/marketplace-operator/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.127112 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.128516 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-utilities/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.131486 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-utilities/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.251606 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.303108 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.304222 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/extract-utilities/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.388824 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r42nm_bc43e84a-00a8-4187-86ef-232d444a3415/registry-server/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.506475 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-utilities/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.661150 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.671719 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.674274 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-utilities/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.877720 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-content/0.log" Nov 25 20:06:35 crc kubenswrapper[4759]: I1125 20:06:35.883393 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/extract-utilities/0.log" Nov 25 20:06:36 crc kubenswrapper[4759]: I1125 20:06:36.241793 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhw6t_4570680c-2f2e-44dc-aeec-caf3dd5231ef/registry-server/0.log" Nov 25 20:07:39 crc kubenswrapper[4759]: I1125 20:07:39.410075 4759 generic.go:334] "Generic (PLEG): container finished" podID="b0772673-9511-4e08-85ca-853958cf8519" containerID="4ce5b86703d3c675bd30827ca984fd55b609fe4d406b649303c9ebc7bb7b164e" exitCode=0 Nov 25 20:07:39 crc kubenswrapper[4759]: I1125 20:07:39.410148 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" event={"ID":"b0772673-9511-4e08-85ca-853958cf8519","Type":"ContainerDied","Data":"4ce5b86703d3c675bd30827ca984fd55b609fe4d406b649303c9ebc7bb7b164e"} Nov 25 20:07:39 crc kubenswrapper[4759]: I1125 20:07:39.411154 4759 scope.go:117] "RemoveContainer" containerID="4ce5b86703d3c675bd30827ca984fd55b609fe4d406b649303c9ebc7bb7b164e" Nov 25 20:07:40 crc kubenswrapper[4759]: I1125 20:07:40.236346 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s6tfq_must-gather-5hrqv_b0772673-9511-4e08-85ca-853958cf8519/gather/0.log" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.205236 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s6tfq/must-gather-5hrqv"] Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.206672 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="copy" containerID="cri-o://3ed8d47158d6e1a28ec3fbfe1f6b9ec84799fc9c3a90e61c97871ca2ce2f97da" gracePeriod=2 Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.211028 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s6tfq/must-gather-5hrqv"] Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.486775 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s6tfq_must-gather-5hrqv_b0772673-9511-4e08-85ca-853958cf8519/copy/0.log" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.487379 4759 generic.go:334] "Generic (PLEG): container finished" podID="b0772673-9511-4e08-85ca-853958cf8519" containerID="3ed8d47158d6e1a28ec3fbfe1f6b9ec84799fc9c3a90e61c97871ca2ce2f97da" exitCode=143 Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.597721 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s6tfq_must-gather-5hrqv_b0772673-9511-4e08-85ca-853958cf8519/copy/0.log" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.598072 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.703922 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjgb8\" (UniqueName: \"kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8\") pod \"b0772673-9511-4e08-85ca-853958cf8519\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.703988 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output\") pod \"b0772673-9511-4e08-85ca-853958cf8519\" (UID: \"b0772673-9511-4e08-85ca-853958cf8519\") " Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.712637 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8" (OuterVolumeSpecName: "kube-api-access-cjgb8") pod "b0772673-9511-4e08-85ca-853958cf8519" (UID: "b0772673-9511-4e08-85ca-853958cf8519"). InnerVolumeSpecName "kube-api-access-cjgb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.773865 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b0772673-9511-4e08-85ca-853958cf8519" (UID: "b0772673-9511-4e08-85ca-853958cf8519"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.806063 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjgb8\" (UniqueName: \"kubernetes.io/projected/b0772673-9511-4e08-85ca-853958cf8519-kube-api-access-cjgb8\") on node \"crc\" DevicePath \"\"" Nov 25 20:07:47 crc kubenswrapper[4759]: I1125 20:07:47.806104 4759 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0772673-9511-4e08-85ca-853958cf8519-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 20:07:48 crc kubenswrapper[4759]: I1125 20:07:48.114955 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0772673-9511-4e08-85ca-853958cf8519" path="/var/lib/kubelet/pods/b0772673-9511-4e08-85ca-853958cf8519/volumes" Nov 25 20:07:48 crc kubenswrapper[4759]: I1125 20:07:48.495054 4759 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s6tfq_must-gather-5hrqv_b0772673-9511-4e08-85ca-853958cf8519/copy/0.log" Nov 25 20:07:48 crc kubenswrapper[4759]: I1125 20:07:48.495402 4759 scope.go:117] "RemoveContainer" containerID="3ed8d47158d6e1a28ec3fbfe1f6b9ec84799fc9c3a90e61c97871ca2ce2f97da" Nov 25 20:07:48 crc kubenswrapper[4759]: I1125 20:07:48.495423 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s6tfq/must-gather-5hrqv" Nov 25 20:07:48 crc kubenswrapper[4759]: I1125 20:07:48.517354 4759 scope.go:117] "RemoveContainer" containerID="4ce5b86703d3c675bd30827ca984fd55b609fe4d406b649303c9ebc7bb7b164e" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.327195 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:03 crc kubenswrapper[4759]: E1125 20:08:03.327914 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="gather" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.327925 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="gather" Nov 25 20:08:03 crc kubenswrapper[4759]: E1125 20:08:03.327959 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="copy" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.327965 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="copy" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.328078 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="gather" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.328093 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0772673-9511-4e08-85ca-853958cf8519" containerName="copy" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.329024 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.340058 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.463570 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.463919 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.463998 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v9qh\" (UniqueName: \"kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.565807 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.565861 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.565940 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v9qh\" (UniqueName: \"kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.566262 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.566778 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.588422 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v9qh\" (UniqueName: \"kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh\") pod \"certified-operators-n2rfm\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:03 crc kubenswrapper[4759]: I1125 20:08:03.666708 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:04 crc kubenswrapper[4759]: I1125 20:08:04.088696 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:04 crc kubenswrapper[4759]: I1125 20:08:04.636918 4759 generic.go:334] "Generic (PLEG): container finished" podID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerID="c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de" exitCode=0 Nov 25 20:08:04 crc kubenswrapper[4759]: I1125 20:08:04.636991 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerDied","Data":"c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de"} Nov 25 20:08:04 crc kubenswrapper[4759]: I1125 20:08:04.637289 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerStarted","Data":"37e7ffa1fab8b8440ffe78676c9db463569310af34fa99a6bd426f8f6bfa9512"} Nov 25 20:08:05 crc kubenswrapper[4759]: I1125 20:08:05.644649 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerStarted","Data":"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822"} Nov 25 20:08:06 crc kubenswrapper[4759]: I1125 20:08:06.653281 4759 generic.go:334] "Generic (PLEG): container finished" podID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerID="845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822" exitCode=0 Nov 25 20:08:06 crc kubenswrapper[4759]: I1125 20:08:06.653330 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerDied","Data":"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822"} Nov 25 20:08:07 crc kubenswrapper[4759]: I1125 20:08:07.661831 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerStarted","Data":"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5"} Nov 25 20:08:07 crc kubenswrapper[4759]: I1125 20:08:07.683798 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n2rfm" podStartSLOduration=2.250614949 podStartE2EDuration="4.683781468s" podCreationTimestamp="2025-11-25 20:08:03 +0000 UTC" firstStartedPulling="2025-11-25 20:08:04.640088811 +0000 UTC m=+2075.226036512" lastFinishedPulling="2025-11-25 20:08:07.07325532 +0000 UTC m=+2077.659203031" observedRunningTime="2025-11-25 20:08:07.678519347 +0000 UTC m=+2078.264467058" watchObservedRunningTime="2025-11-25 20:08:07.683781468 +0000 UTC m=+2078.269729169" Nov 25 20:08:13 crc kubenswrapper[4759]: I1125 20:08:13.667932 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:13 crc kubenswrapper[4759]: I1125 20:08:13.668565 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:13 crc kubenswrapper[4759]: I1125 20:08:13.714980 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:13 crc kubenswrapper[4759]: I1125 20:08:13.764126 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.325514 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.326325 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n2rfm" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="registry-server" containerID="cri-o://3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5" gracePeriod=2 Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.723103 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.777180 4759 generic.go:334] "Generic (PLEG): container finished" podID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerID="3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5" exitCode=0 Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.777269 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerDied","Data":"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5"} Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.777304 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2rfm" event={"ID":"dfdf60ae-49cf-4978-b412-b3c4341aa26b","Type":"ContainerDied","Data":"37e7ffa1fab8b8440ffe78676c9db463569310af34fa99a6bd426f8f6bfa9512"} Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.777339 4759 scope.go:117] "RemoveContainer" containerID="3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.777704 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2rfm" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.794920 4759 scope.go:117] "RemoveContainer" containerID="845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.810192 4759 scope.go:117] "RemoveContainer" containerID="c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.839564 4759 scope.go:117] "RemoveContainer" containerID="3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5" Nov 25 20:08:22 crc kubenswrapper[4759]: E1125 20:08:22.840016 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5\": container with ID starting with 3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5 not found: ID does not exist" containerID="3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.840053 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5"} err="failed to get container status \"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5\": rpc error: code = NotFound desc = could not find container \"3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5\": container with ID starting with 3893e4ceaa92110ddf10af518076c976940ba543aa84fffd4ee429bbd9555ab5 not found: ID does not exist" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.840077 4759 scope.go:117] "RemoveContainer" containerID="845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822" Nov 25 20:08:22 crc kubenswrapper[4759]: E1125 20:08:22.840358 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822\": container with ID starting with 845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822 not found: ID does not exist" containerID="845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.840385 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822"} err="failed to get container status \"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822\": rpc error: code = NotFound desc = could not find container \"845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822\": container with ID starting with 845f9d3d8255112d3c13bffe0369a2b85dde70f28cb90f24bd682476b9d68822 not found: ID does not exist" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.840403 4759 scope.go:117] "RemoveContainer" containerID="c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de" Nov 25 20:08:22 crc kubenswrapper[4759]: E1125 20:08:22.840720 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de\": container with ID starting with c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de not found: ID does not exist" containerID="c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.840749 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de"} err="failed to get container status \"c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de\": rpc error: code = NotFound desc = could not find container \"c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de\": container with ID starting with c0484c0d6de4ed680bb8be6988a124bcc7adc0cac09fa3cd51860bb8709367de not found: ID does not exist" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.868419 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v9qh\" (UniqueName: \"kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh\") pod \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.868519 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities\") pod \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.868600 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content\") pod \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\" (UID: \"dfdf60ae-49cf-4978-b412-b3c4341aa26b\") " Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.869615 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities" (OuterVolumeSpecName: "utilities") pod "dfdf60ae-49cf-4978-b412-b3c4341aa26b" (UID: "dfdf60ae-49cf-4978-b412-b3c4341aa26b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.874034 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh" (OuterVolumeSpecName: "kube-api-access-6v9qh") pod "dfdf60ae-49cf-4978-b412-b3c4341aa26b" (UID: "dfdf60ae-49cf-4978-b412-b3c4341aa26b"). InnerVolumeSpecName "kube-api-access-6v9qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.918372 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfdf60ae-49cf-4978-b412-b3c4341aa26b" (UID: "dfdf60ae-49cf-4978-b412-b3c4341aa26b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.970394 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v9qh\" (UniqueName: \"kubernetes.io/projected/dfdf60ae-49cf-4978-b412-b3c4341aa26b-kube-api-access-6v9qh\") on node \"crc\" DevicePath \"\"" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.970427 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 20:08:22 crc kubenswrapper[4759]: I1125 20:08:22.970439 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfdf60ae-49cf-4978-b412-b3c4341aa26b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 20:08:23 crc kubenswrapper[4759]: I1125 20:08:23.110180 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:23 crc kubenswrapper[4759]: I1125 20:08:23.114573 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n2rfm"] Nov 25 20:08:24 crc kubenswrapper[4759]: I1125 20:08:24.118317 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" path="/var/lib/kubelet/pods/dfdf60ae-49cf-4978-b412-b3c4341aa26b/volumes" Nov 25 20:08:54 crc kubenswrapper[4759]: I1125 20:08:54.895273 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:08:54 crc kubenswrapper[4759]: I1125 20:08:54.895847 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:09:24 crc kubenswrapper[4759]: I1125 20:09:24.896150 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:09:24 crc kubenswrapper[4759]: I1125 20:09:24.897739 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:09:54 crc kubenswrapper[4759]: I1125 20:09:54.895885 4759 patch_prober.go:28] interesting pod/machine-config-daemon-rfsr5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 20:09:54 crc kubenswrapper[4759]: I1125 20:09:54.896421 4759 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 20:09:54 crc kubenswrapper[4759]: I1125 20:09:54.896485 4759 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" Nov 25 20:09:54 crc kubenswrapper[4759]: I1125 20:09:54.897130 4759 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34527af12c7198aeb773451b05c7977f1f065c3b0959196146823ccb0c85db94"} pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 20:09:54 crc kubenswrapper[4759]: I1125 20:09:54.897172 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" podUID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerName="machine-config-daemon" containerID="cri-o://34527af12c7198aeb773451b05c7977f1f065c3b0959196146823ccb0c85db94" gracePeriod=600 Nov 25 20:09:55 crc kubenswrapper[4759]: I1125 20:09:55.444333 4759 generic.go:334] "Generic (PLEG): container finished" podID="07c1b558-ca9e-4edc-9420-05750ceee3a9" containerID="34527af12c7198aeb773451b05c7977f1f065c3b0959196146823ccb0c85db94" exitCode=0 Nov 25 20:09:55 crc kubenswrapper[4759]: I1125 20:09:55.444370 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerDied","Data":"34527af12c7198aeb773451b05c7977f1f065c3b0959196146823ccb0c85db94"} Nov 25 20:09:55 crc kubenswrapper[4759]: I1125 20:09:55.444699 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfsr5" event={"ID":"07c1b558-ca9e-4edc-9420-05750ceee3a9","Type":"ContainerStarted","Data":"cb2edded5a34031447b0ec8a0ebc9e4b7b14a130594f4ae2a60dfe884362bfad"} Nov 25 20:09:55 crc kubenswrapper[4759]: I1125 20:09:55.444716 4759 scope.go:117] "RemoveContainer" containerID="55c90e0a98cd332c3c0a78ef3b4794d9c4228fc18ef9e6a7555368eda5ba78ad" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.376668 4759 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4mgxh"] Nov 25 20:10:08 crc kubenswrapper[4759]: E1125 20:10:08.377427 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="extract-utilities" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.377438 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="extract-utilities" Nov 25 20:10:08 crc kubenswrapper[4759]: E1125 20:10:08.377485 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="extract-content" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.377491 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="extract-content" Nov 25 20:10:08 crc kubenswrapper[4759]: E1125 20:10:08.377502 4759 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="registry-server" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.377508 4759 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="registry-server" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.377620 4759 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfdf60ae-49cf-4978-b412-b3c4341aa26b" containerName="registry-server" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.378535 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.404174 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mgxh"] Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.522009 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-utilities\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.522062 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-catalog-content\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.522127 4759 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64gjv\" (UniqueName: \"kubernetes.io/projected/5b408796-2db9-427b-9da4-62f77e1ba1ea-kube-api-access-64gjv\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.624576 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-utilities\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.624647 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-catalog-content\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.624704 4759 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64gjv\" (UniqueName: \"kubernetes.io/projected/5b408796-2db9-427b-9da4-62f77e1ba1ea-kube-api-access-64gjv\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.625200 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-utilities\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.625322 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b408796-2db9-427b-9da4-62f77e1ba1ea-catalog-content\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.645195 4759 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64gjv\" (UniqueName: \"kubernetes.io/projected/5b408796-2db9-427b-9da4-62f77e1ba1ea-kube-api-access-64gjv\") pod \"community-operators-4mgxh\" (UID: \"5b408796-2db9-427b-9da4-62f77e1ba1ea\") " pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:08 crc kubenswrapper[4759]: I1125 20:10:08.699888 4759 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:09 crc kubenswrapper[4759]: I1125 20:10:09.228811 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mgxh"] Nov 25 20:10:09 crc kubenswrapper[4759]: I1125 20:10:09.549498 4759 generic.go:334] "Generic (PLEG): container finished" podID="5b408796-2db9-427b-9da4-62f77e1ba1ea" containerID="174f7f6e3af87e61f5ac7e6471b64419f91a6786476147bbdbe772abbbea9bf8" exitCode=0 Nov 25 20:10:09 crc kubenswrapper[4759]: I1125 20:10:09.549547 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mgxh" event={"ID":"5b408796-2db9-427b-9da4-62f77e1ba1ea","Type":"ContainerDied","Data":"174f7f6e3af87e61f5ac7e6471b64419f91a6786476147bbdbe772abbbea9bf8"} Nov 25 20:10:09 crc kubenswrapper[4759]: I1125 20:10:09.549802 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mgxh" event={"ID":"5b408796-2db9-427b-9da4-62f77e1ba1ea","Type":"ContainerStarted","Data":"e767a17c6f49a7af05b655158f67cb32500c0efcb041bcbff890570b782b0045"} Nov 25 20:10:09 crc kubenswrapper[4759]: I1125 20:10:09.552067 4759 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 20:10:13 crc kubenswrapper[4759]: I1125 20:10:13.576074 4759 generic.go:334] "Generic (PLEG): container finished" podID="5b408796-2db9-427b-9da4-62f77e1ba1ea" containerID="454b11f23993ec55ab79f60a95f208c319a7d297d7f099bb8253f9dafafcac51" exitCode=0 Nov 25 20:10:13 crc kubenswrapper[4759]: I1125 20:10:13.576186 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mgxh" event={"ID":"5b408796-2db9-427b-9da4-62f77e1ba1ea","Type":"ContainerDied","Data":"454b11f23993ec55ab79f60a95f208c319a7d297d7f099bb8253f9dafafcac51"} Nov 25 20:10:13 crc kubenswrapper[4759]: E1125 20:10:13.635070 4759 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b408796_2db9_427b_9da4_62f77e1ba1ea.slice/crio-conmon-454b11f23993ec55ab79f60a95f208c319a7d297d7f099bb8253f9dafafcac51.scope\": RecentStats: unable to find data in memory cache]" Nov 25 20:10:14 crc kubenswrapper[4759]: I1125 20:10:14.585426 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mgxh" event={"ID":"5b408796-2db9-427b-9da4-62f77e1ba1ea","Type":"ContainerStarted","Data":"a82ae95e77543624701ad5a8fdf389ddd5ff9c8a5aa1efecad3904d2bfe6b3e4"} Nov 25 20:10:14 crc kubenswrapper[4759]: I1125 20:10:14.606460 4759 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4mgxh" podStartSLOduration=2.092766681 podStartE2EDuration="6.606427957s" podCreationTimestamp="2025-11-25 20:10:08 +0000 UTC" firstStartedPulling="2025-11-25 20:10:09.551681332 +0000 UTC m=+2200.137629033" lastFinishedPulling="2025-11-25 20:10:14.065342608 +0000 UTC m=+2204.651290309" observedRunningTime="2025-11-25 20:10:14.599348047 +0000 UTC m=+2205.185295758" watchObservedRunningTime="2025-11-25 20:10:14.606427957 +0000 UTC m=+2205.192375648" Nov 25 20:10:18 crc kubenswrapper[4759]: I1125 20:10:18.700615 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:18 crc kubenswrapper[4759]: I1125 20:10:18.701090 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:18 crc kubenswrapper[4759]: I1125 20:10:18.765677 4759 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:19 crc kubenswrapper[4759]: I1125 20:10:19.719089 4759 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4mgxh" Nov 25 20:10:19 crc kubenswrapper[4759]: I1125 20:10:19.806477 4759 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mgxh"] Nov 25 20:10:19 crc kubenswrapper[4759]: I1125 20:10:19.847768 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 20:10:19 crc kubenswrapper[4759]: I1125 20:10:19.848423 4759 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5scxm" podUID="00a8d174-0d8b-4841-ace2-958422c2d299" containerName="registry-server" containerID="cri-o://2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf" gracePeriod=2 Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.314431 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5scxm" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.402111 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities\") pod \"00a8d174-0d8b-4841-ace2-958422c2d299\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.402185 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content\") pod \"00a8d174-0d8b-4841-ace2-958422c2d299\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.402224 4759 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vljtv\" (UniqueName: \"kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv\") pod \"00a8d174-0d8b-4841-ace2-958422c2d299\" (UID: \"00a8d174-0d8b-4841-ace2-958422c2d299\") " Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.412045 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities" (OuterVolumeSpecName: "utilities") pod "00a8d174-0d8b-4841-ace2-958422c2d299" (UID: "00a8d174-0d8b-4841-ace2-958422c2d299"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.414621 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv" (OuterVolumeSpecName: "kube-api-access-vljtv") pod "00a8d174-0d8b-4841-ace2-958422c2d299" (UID: "00a8d174-0d8b-4841-ace2-958422c2d299"). InnerVolumeSpecName "kube-api-access-vljtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.464459 4759 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00a8d174-0d8b-4841-ace2-958422c2d299" (UID: "00a8d174-0d8b-4841-ace2-958422c2d299"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.503547 4759 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.503586 4759 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a8d174-0d8b-4841-ace2-958422c2d299-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.503601 4759 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vljtv\" (UniqueName: \"kubernetes.io/projected/00a8d174-0d8b-4841-ace2-958422c2d299-kube-api-access-vljtv\") on node \"crc\" DevicePath \"\"" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.687511 4759 generic.go:334] "Generic (PLEG): container finished" podID="00a8d174-0d8b-4841-ace2-958422c2d299" containerID="2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf" exitCode=0 Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.688509 4759 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5scxm" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.688826 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerDied","Data":"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf"} Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.688885 4759 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5scxm" event={"ID":"00a8d174-0d8b-4841-ace2-958422c2d299","Type":"ContainerDied","Data":"fd194fd6ea15f3c70f232ce15673790ce372b8ccd4dd1012672a7429d75bc287"} Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.688909 4759 scope.go:117] "RemoveContainer" containerID="2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.715397 4759 scope.go:117] "RemoveContainer" containerID="2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.735562 4759 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.746279 4759 scope.go:117] "RemoveContainer" containerID="73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.748726 4759 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5scxm"] Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.791847 4759 scope.go:117] "RemoveContainer" containerID="2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf" Nov 25 20:10:20 crc kubenswrapper[4759]: E1125 20:10:20.793691 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf\": container with ID starting with 2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf not found: ID does not exist" containerID="2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.793730 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf"} err="failed to get container status \"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf\": rpc error: code = NotFound desc = could not find container \"2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf\": container with ID starting with 2aa1fd43d2a03c6287981350c87ee968a391eae8650a9f59060a1bdd8f5e7faf not found: ID does not exist" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.793757 4759 scope.go:117] "RemoveContainer" containerID="2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7" Nov 25 20:10:20 crc kubenswrapper[4759]: E1125 20:10:20.794946 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7\": container with ID starting with 2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7 not found: ID does not exist" containerID="2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.794986 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7"} err="failed to get container status \"2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7\": rpc error: code = NotFound desc = could not find container \"2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7\": container with ID starting with 2143dfd0dc47fb1d1c68d656d085d4e09278c558e2c593c25e1e263406319fe7 not found: ID does not exist" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.795015 4759 scope.go:117] "RemoveContainer" containerID="73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f" Nov 25 20:10:20 crc kubenswrapper[4759]: E1125 20:10:20.795342 4759 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f\": container with ID starting with 73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f not found: ID does not exist" containerID="73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f" Nov 25 20:10:20 crc kubenswrapper[4759]: I1125 20:10:20.795380 4759 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f"} err="failed to get container status \"73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f\": rpc error: code = NotFound desc = could not find container \"73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f\": container with ID starting with 73bb1ba38a561061917ff1183c90b3924728df5363c4039faa165214074bc33f not found: ID does not exist" Nov 25 20:10:22 crc kubenswrapper[4759]: I1125 20:10:22.118843 4759 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a8d174-0d8b-4841-ace2-958422c2d299" path="/var/lib/kubelet/pods/00a8d174-0d8b-4841-ace2-958422c2d299/volumes" Nov 25 20:10:31 crc kubenswrapper[4759]: I1125 20:10:31.889437 4759 scope.go:117] "RemoveContainer" containerID="cec487aa8556481a16785014f489b6e2fc841d6d78aa94f67a063235963404be" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111406620024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111406621017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111401653016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111401654015453 5ustar corecore